diff --git a/.gitattributes b/.gitattributes index 55e3a15d0b58cf70db47b29f4f36992187827302..45e9d12987334d97ac6c0cbb1888fe97db702feb 100644 --- a/.gitattributes +++ b/.gitattributes @@ -63,3 +63,14 @@ runs/test-model/checkpoint-160/tokenizer.json filter=lfs diff=lfs merge=lfs -tex runs/test-model/checkpoint-240/tokenizer.json filter=lfs diff=lfs merge=lfs -text runs/test-model/checkpoint-320/tokenizer.json filter=lfs diff=lfs merge=lfs -text runs/test-model/checkpoint-80/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-1080/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-1440/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-1800/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-2160/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-2520/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-2880/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-3240/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-360/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-3600/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-72/tokenizer.json filter=lfs diff=lfs merge=lfs -text +runs/baseline/checkpoint-720/tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/runs/baseline/checkpoint-1080/chat_template.jinja b/runs/baseline/checkpoint-1080/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-1080/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-1080/config.json b/runs/baseline/checkpoint-1080/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-1080/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-1080/generation_config.json b/runs/baseline/checkpoint-1080/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-1080/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-1080/model.safetensors b/runs/baseline/checkpoint-1080/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..af8dce58d95b2bf66f36168d7a18ed70fbf1f4df --- /dev/null +++ b/runs/baseline/checkpoint-1080/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd34eed88b2d67d28b6f780a8191a0103043d184014cd0901149410e227b1a76 +size 583356232 diff --git a/runs/baseline/checkpoint-1080/optimizer.pt b/runs/baseline/checkpoint-1080/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..1cb51be1ecbe79e6e68f3b273a495105475d6df2 --- /dev/null +++ b/runs/baseline/checkpoint-1080/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:838098c21ee71b236879956ab772f05931eed8de49d4db467033e5966af7ca33 +size 1166825803 diff --git a/runs/baseline/checkpoint-1080/rng_state_0.pth b/runs/baseline/checkpoint-1080/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..ff81a4a04de793f1c690eab304502628a6f01a16 --- /dev/null +++ b/runs/baseline/checkpoint-1080/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e64ff3db32365cabac6ffad743c22ab2586a3c7538842f851d54fcc9b00320b +size 14917 diff --git a/runs/baseline/checkpoint-1080/rng_state_1.pth b/runs/baseline/checkpoint-1080/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..2e0fc04a0aae1c0cb2027f4e52b0d2e25910d9a8 --- /dev/null +++ b/runs/baseline/checkpoint-1080/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:837b65fefb8e1f24a6981b0d4649ff164abbce25b5c9c0d885ed58d7e0986bae +size 14917 diff --git a/runs/baseline/checkpoint-1080/scheduler.pt b/runs/baseline/checkpoint-1080/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..8736048f38a59bc77b7ec1746ae478e02d0b7c5c --- /dev/null +++ b/runs/baseline/checkpoint-1080/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0e299effdfaa8cee7008c9e7d1521bed82d9b09e5041046b2ae409f4f1f8149 +size 1465 diff --git a/runs/baseline/checkpoint-1080/tokenizer.json b/runs/baseline/checkpoint-1080/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-1080/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-1080/tokenizer_config.json b/runs/baseline/checkpoint-1080/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-1080/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-1080/trainer_state.json b/runs/baseline/checkpoint-1080/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..595972e0b0591e4742ea62dd80b78a9ea52fd474 --- /dev/null +++ b/runs/baseline/checkpoint-1080/trainer_state.json @@ -0,0 +1,7706 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 15.0, + "eval_steps": 500, + "global_step": 1080, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4.604420041998336e+17, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-1080/training_args.bin b/runs/baseline/checkpoint-1080/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-1080/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-1440/chat_template.jinja b/runs/baseline/checkpoint-1440/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-1440/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-1440/config.json b/runs/baseline/checkpoint-1440/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-1440/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-1440/generation_config.json b/runs/baseline/checkpoint-1440/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-1440/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-1440/model.safetensors b/runs/baseline/checkpoint-1440/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6365dbd1359578638aa2e7a4cd5e47df42b85c11 --- /dev/null +++ b/runs/baseline/checkpoint-1440/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d590f62abdc2e7d7be25aacb35cb7e0eb2e858370679a61e7caec37105ed0d5 +size 583356232 diff --git a/runs/baseline/checkpoint-1440/optimizer.pt b/runs/baseline/checkpoint-1440/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b80ec4b8098716014384f3d375b6e2f58c238f9b --- /dev/null +++ b/runs/baseline/checkpoint-1440/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bf4afa93e6537e6e0f119f6af49e6a223f95a084b81e99ed26287e4467cd6a7 +size 1166825803 diff --git a/runs/baseline/checkpoint-1440/rng_state_0.pth b/runs/baseline/checkpoint-1440/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..3c396253e5ad58aa3254e1cb9364198815e3ba61 --- /dev/null +++ b/runs/baseline/checkpoint-1440/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad41999a3021e51cc92b222472cf4756a254132cabf58c4c2e1ee206c7366885 +size 14917 diff --git a/runs/baseline/checkpoint-1440/rng_state_1.pth b/runs/baseline/checkpoint-1440/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..08b5f24cb1ec06dc192e471442f86bfe5e398030 --- /dev/null +++ b/runs/baseline/checkpoint-1440/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29848a162fabe62f2066ebb397d9dc38254fb7b578fd1c8908975df46738482d +size 14917 diff --git a/runs/baseline/checkpoint-1440/scheduler.pt b/runs/baseline/checkpoint-1440/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..032c2c72b5314d57cafe9ae23600e7369414b19b --- /dev/null +++ b/runs/baseline/checkpoint-1440/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc198898504ac73bd5b7b2d8d7dbdfa658d335ae2db41ad71022be987a114f4e +size 1465 diff --git a/runs/baseline/checkpoint-1440/tokenizer.json b/runs/baseline/checkpoint-1440/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-1440/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-1440/tokenizer_config.json b/runs/baseline/checkpoint-1440/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-1440/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-1440/trainer_state.json b/runs/baseline/checkpoint-1440/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..1f5b852ec31bac36850963888bdb68810449cb10 --- /dev/null +++ b/runs/baseline/checkpoint-1440/trainer_state.json @@ -0,0 +1,10266 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 20.0, + "eval_steps": 500, + "global_step": 1440, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 6.139226722664448e+17, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-1440/training_args.bin b/runs/baseline/checkpoint-1440/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-1440/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-1800/chat_template.jinja b/runs/baseline/checkpoint-1800/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-1800/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-1800/config.json b/runs/baseline/checkpoint-1800/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-1800/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-1800/generation_config.json b/runs/baseline/checkpoint-1800/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-1800/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-1800/model.safetensors b/runs/baseline/checkpoint-1800/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8327df03d22e26408ccf723ec1ca3f534538fa36 --- /dev/null +++ b/runs/baseline/checkpoint-1800/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c2ac76aaab232d5b01beedf4207ba5eea5d85fe47eae4f88f5cf93160abb37d +size 583356232 diff --git a/runs/baseline/checkpoint-1800/optimizer.pt b/runs/baseline/checkpoint-1800/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..7448dcb39d76879aae6cff91ec20c9f64cbd77a0 --- /dev/null +++ b/runs/baseline/checkpoint-1800/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13614eed9a3624e21a4f90498242d9bb1c4283ed2a4e2786e604d0a083b1cb5a +size 1166825803 diff --git a/runs/baseline/checkpoint-1800/rng_state_0.pth b/runs/baseline/checkpoint-1800/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..304a7576d9bd060137258dba420a3e1ff00587c1 --- /dev/null +++ b/runs/baseline/checkpoint-1800/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:987694c44f0af3a8ae4e43e3677bef48a1cfa8e57eee3832ef223e9b8ad80516 +size 14917 diff --git a/runs/baseline/checkpoint-1800/rng_state_1.pth b/runs/baseline/checkpoint-1800/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..75486c47bfa9499bfe91975e3af3355c9526365c --- /dev/null +++ b/runs/baseline/checkpoint-1800/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea5ecba78b648c2a45cab91d590e52b7295077372e6bf91c1591d3fb6a5535ca +size 14917 diff --git a/runs/baseline/checkpoint-1800/scheduler.pt b/runs/baseline/checkpoint-1800/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..7bf0bbb2d908712e094e342325183f9d764f7a64 --- /dev/null +++ b/runs/baseline/checkpoint-1800/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1662fc1ce9587c318535917c43a6b1f1e5532c3833011159b3431966308a9b2 +size 1465 diff --git a/runs/baseline/checkpoint-1800/tokenizer.json b/runs/baseline/checkpoint-1800/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-1800/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-1800/tokenizer_config.json b/runs/baseline/checkpoint-1800/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-1800/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-1800/trainer_state.json b/runs/baseline/checkpoint-1800/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b4b06705b9f0abc5b4f2960828a410f5a9e4c9c7 --- /dev/null +++ b/runs/baseline/checkpoint-1800/trainer_state.json @@ -0,0 +1,12826 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 25.0, + "eval_steps": 500, + "global_step": 1800, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.641542434692383, + "eval_runtime": 51.6342, + "eval_samples_per_second": 47.294, + "eval_steps_per_second": 1.491, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.009694516658782959, + "learning_rate": 0.0020492498088767836, + "loss": 4.32808256149292, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.0097952950745821, + "learning_rate": 0.002048005909777114, + "loss": 4.280096054077148, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.010393359698355198, + "learning_rate": 0.0020467615812442106, + "loss": 4.317824363708496, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.00977372657507658, + "learning_rate": 0.0020455168242806054, + "loss": 4.3445329666137695, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.008568719029426575, + "learning_rate": 0.0020442716398891775, + "loss": 4.318026542663574, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.009732510894536972, + "learning_rate": 0.0020430260290731487, + "loss": 4.330609321594238, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.010429969057440758, + "learning_rate": 0.0020417799928360864, + "loss": 4.316794395446777, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.009292383678257465, + "learning_rate": 0.002040533532181898, + "loss": 4.317420959472656, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.011222455650568008, + "learning_rate": 0.0020392866481148356, + "loss": 4.292610168457031, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.013591725379228592, + "learning_rate": 0.0020380393416394894, + "loss": 4.325689315795898, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.013572026044130325, + "learning_rate": 0.002036791613760794, + "loss": 4.334960460662842, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.012986521236598492, + "learning_rate": 0.0020355434654840197, + "loss": 4.33829402923584, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.013321424834430218, + "learning_rate": 0.0020342948978147777, + "loss": 4.2926926612854, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.012307215481996536, + "learning_rate": 0.0020330459117590153, + "loss": 4.32757568359375, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.012909360229969025, + "learning_rate": 0.0020317965083230197, + "loss": 4.362889289855957, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017746610566973686, + "learning_rate": 0.0020305466885134115, + "loss": 4.3426313400268555, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.01696564070880413, + "learning_rate": 0.002029296453337148, + "loss": 4.3298516273498535, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.013348255306482315, + "learning_rate": 0.002028045803801522, + "loss": 4.360220432281494, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.013225145637989044, + "learning_rate": 0.002026794740914158, + "loss": 4.346250057220459, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.01569020375609398, + "learning_rate": 0.002025543265683016, + "loss": 4.308025360107422, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.018612321466207504, + "learning_rate": 0.002024291379116385, + "loss": 4.330190181732178, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.01619943603873253, + "learning_rate": 0.00202303908222289, + "loss": 4.339552402496338, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.015385213308036327, + "learning_rate": 0.0020217863760114826, + "loss": 4.3412184715271, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.018517112359404564, + "learning_rate": 0.002020533261491446, + "loss": 4.3289313316345215, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.014868509024381638, + "learning_rate": 0.0020192797396723903, + "loss": 4.359846115112305, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.014240248128771782, + "learning_rate": 0.0020180258115642578, + "loss": 4.369707107543945, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.018113158643245697, + "learning_rate": 0.0020167714781773143, + "loss": 4.362668514251709, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.01802460104227066, + "learning_rate": 0.002015516740522154, + "loss": 4.344742298126221, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.01414549071341753, + "learning_rate": 0.0020142615996096964, + "loss": 4.358080863952637, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.012157700955867767, + "learning_rate": 0.002013006056451186, + "loss": 4.347763538360596, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.011268534697592258, + "learning_rate": 0.0020117501120581905, + "loss": 4.335859298706055, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.010395628400146961, + "learning_rate": 0.002010493767442602, + "loss": 4.341609001159668, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.012147950939834118, + "learning_rate": 0.002009237023616635, + "loss": 4.328502655029297, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.00945813488215208, + "learning_rate": 0.0020079798815928252, + "loss": 4.311186790466309, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.009489500895142555, + "learning_rate": 0.0020067223423840294, + "loss": 4.371470928192139, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.009360029362142086, + "learning_rate": 0.0020054644070034234, + "loss": 4.33941125869751, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.010032779537141323, + "learning_rate": 0.0020042060764645037, + "loss": 4.365797519683838, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.009367046877741814, + "learning_rate": 0.002002947351781084, + "loss": 4.29443359375, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.008369733579456806, + "learning_rate": 0.0020016882339672967, + "loss": 4.294233322143555, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.010394785553216934, + "learning_rate": 0.0020004287240375893, + "loss": 4.352355003356934, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01068816613405943, + "learning_rate": 0.001999168823006728, + "loss": 4.342516899108887, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.009975973516702652, + "learning_rate": 0.0019979085318897905, + "loss": 4.346763610839844, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.009618151932954788, + "learning_rate": 0.001996647851702171, + "loss": 4.329115867614746, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.008399694226682186, + "learning_rate": 0.001995386783459577, + "loss": 4.28736686706543, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.008235268294811249, + "learning_rate": 0.001994125328178029, + "loss": 4.339872360229492, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.00836547277867794, + "learning_rate": 0.0019928634868738577, + "loss": 4.328156471252441, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.009619626216590405, + "learning_rate": 0.001991601260563706, + "loss": 4.337394714355469, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.009814038872718811, + "learning_rate": 0.0019903386502645275, + "loss": 4.325926780700684, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.009588954970240593, + "learning_rate": 0.001989075656993584, + "loss": 4.286477088928223, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.010471943765878677, + "learning_rate": 0.001987812281768447, + "loss": 4.322466850280762, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.010239035822451115, + "learning_rate": 0.001986548525606994, + "loss": 4.3487653732299805, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.008935568854212761, + "learning_rate": 0.001985284389527411, + "loss": 4.303601264953613, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.008624759502708912, + "learning_rate": 0.00198401987454819, + "loss": 4.34036922454834, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.009296547621488571, + "learning_rate": 0.0019827549816881277, + "loss": 4.273357391357422, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.009071138687431812, + "learning_rate": 0.0019814897119663248, + "loss": 4.321051597595215, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.008341389708220959, + "learning_rate": 0.0019802240664021876, + "loss": 4.343242645263672, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.007881715893745422, + "learning_rate": 0.0019789580460154227, + "loss": 4.32067346572876, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.008750553242862225, + "learning_rate": 0.0019776916518260404, + "loss": 4.330263614654541, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.008591460064053535, + "learning_rate": 0.001976424884854351, + "loss": 4.311957359313965, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.008084543980658054, + "learning_rate": 0.001975157746120966, + "loss": 4.331319808959961, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.008036594837903976, + "learning_rate": 0.001973890236646797, + "loss": 4.382656097412109, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.00805772002786398, + "learning_rate": 0.0019726223574530527, + "loss": 4.297247886657715, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.009019420482218266, + "learning_rate": 0.0019713541095612405, + "loss": 4.326526165008545, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.009953883476555347, + "learning_rate": 0.0019700854939931653, + "loss": 4.327356338500977, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.009340492077171803, + "learning_rate": 0.0019688165117709273, + "loss": 4.343993186950684, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.010395721532404423, + "learning_rate": 0.0019675471639169226, + "loss": 4.284934997558594, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.009946230798959732, + "learning_rate": 0.0019662774514538415, + "loss": 4.288936614990234, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.010981100611388683, + "learning_rate": 0.0019650073754046695, + "loss": 4.339935779571533, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.009688783437013626, + "learning_rate": 0.0019637369367926827, + "loss": 4.381425857543945, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.009442255832254887, + "learning_rate": 0.001962466136641452, + "loss": 4.304185390472412, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.009371291846036911, + "learning_rate": 0.0019611949759748364, + "loss": 4.306120872497559, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.009544814005494118, + "learning_rate": 0.001959923455816988, + "loss": 4.326251983642578, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.602745532989502, + "eval_runtime": 53.5041, + "eval_samples_per_second": 45.641, + "eval_steps_per_second": 1.439, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.01010364294052124, + "learning_rate": 0.0019586515771923474, + "loss": 4.24491024017334, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.012722729705274105, + "learning_rate": 0.0019573793411256446, + "loss": 4.257739543914795, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.01307204831391573, + "learning_rate": 0.0019561067486418967, + "loss": 4.291684150695801, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.011215966194868088, + "learning_rate": 0.0019548338007664098, + "loss": 4.237947463989258, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.012943019159138203, + "learning_rate": 0.0019535604985247727, + "loss": 4.242859363555908, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.01337282732129097, + "learning_rate": 0.0019522868429428636, + "loss": 4.238934516906738, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.013697991147637367, + "learning_rate": 0.0019510128350468438, + "loss": 4.285760402679443, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.012182159349322319, + "learning_rate": 0.001949738475863158, + "loss": 4.290044784545898, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.012620339170098305, + "learning_rate": 0.0019484637664185346, + "loss": 4.267187118530273, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.011284958571195602, + "learning_rate": 0.0019471887077399837, + "loss": 4.26662540435791, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.01161450706422329, + "learning_rate": 0.001945913300854797, + "loss": 4.283192157745361, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.010022968985140324, + "learning_rate": 0.001944637546790548, + "loss": 4.295843124389648, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.010341878980398178, + "learning_rate": 0.001943361446575087, + "loss": 4.295636177062988, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.011549185030162334, + "learning_rate": 0.0019420850012365457, + "loss": 4.288768768310547, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.012518810108304024, + "learning_rate": 0.0019408082118033331, + "loss": 4.229338645935059, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.012035328894853592, + "learning_rate": 0.0019395310793041353, + "loss": 4.258639335632324, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.011193782091140747, + "learning_rate": 0.0019382536047679146, + "loss": 4.239264488220215, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.009930136613547802, + "learning_rate": 0.0019369757892239096, + "loss": 4.252286434173584, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.00953229982405901, + "learning_rate": 0.0019356976337016327, + "loss": 4.293435096740723, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.010618340224027634, + "learning_rate": 0.0019344191392308716, + "loss": 4.235588550567627, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.008806651458144188, + "learning_rate": 0.0019331403068416856, + "loss": 4.26716423034668, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.008934796787798405, + "learning_rate": 0.0019318611375644065, + "loss": 4.258462905883789, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.009283283725380898, + "learning_rate": 0.0019305816324296386, + "loss": 4.257868766784668, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.009224812500178814, + "learning_rate": 0.0019293017924682555, + "loss": 4.258486747741699, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.008957969956099987, + "learning_rate": 0.0019280216187114016, + "loss": 4.309563636779785, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.010284051299095154, + "learning_rate": 0.0019267411121904894, + "loss": 4.275271415710449, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.009167762473225594, + "learning_rate": 0.0019254602739372001, + "loss": 4.238452911376953, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.009411636739969254, + "learning_rate": 0.0019241791049834816, + "loss": 4.2735595703125, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.009044171310961246, + "learning_rate": 0.0019228976063615497, + "loss": 4.27017879486084, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.008840068243443966, + "learning_rate": 0.0019216157791038834, + "loss": 4.28773307800293, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.009512359276413918, + "learning_rate": 0.001920333624243228, + "loss": 4.266008377075195, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.009946667589247227, + "learning_rate": 0.001919051142812593, + "loss": 4.279062271118164, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.00899894442409277, + "learning_rate": 0.0019177683358452505, + "loss": 4.277679443359375, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.00909668579697609, + "learning_rate": 0.0019164852043747338, + "loss": 4.311397552490234, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.009787327609956264, + "learning_rate": 0.0019152017494348408, + "loss": 4.30875825881958, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.008983899839222431, + "learning_rate": 0.001913917972059626, + "loss": 4.259649276733398, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.008411582559347153, + "learning_rate": 0.0019126338732834071, + "loss": 4.259811878204346, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.009162918664515018, + "learning_rate": 0.0019113494541407579, + "loss": 4.271515846252441, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.011304561980068684, + "learning_rate": 0.0019100647156665125, + "loss": 4.269217491149902, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.012483394704759121, + "learning_rate": 0.0019087796588957617, + "loss": 4.2949299812316895, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.012091380544006824, + "learning_rate": 0.0019074942848638522, + "loss": 4.289752006530762, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.010011567734181881, + "learning_rate": 0.0019062085946063869, + "loss": 4.294095993041992, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.008766217157244682, + "learning_rate": 0.0019049225891592226, + "loss": 4.256425857543945, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.012862488627433777, + "learning_rate": 0.001903636269558471, + "loss": 4.287847995758057, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.013917063362896442, + "learning_rate": 0.0019023496368404972, + "loss": 4.282784938812256, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.012129023671150208, + "learning_rate": 0.0019010626920419173, + "loss": 4.2486419677734375, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.010348426178097725, + "learning_rate": 0.0018997754361995986, + "loss": 4.289836883544922, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.011821693740785122, + "learning_rate": 0.001898487870350662, + "loss": 4.214344024658203, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.013136724010109901, + "learning_rate": 0.0018971999955324742, + "loss": 4.24208402633667, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.013173168525099754, + "learning_rate": 0.0018959118127826538, + "loss": 4.270164489746094, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.011018159799277782, + "learning_rate": 0.0018946233231390656, + "loss": 4.2571916580200195, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.01018658559769392, + "learning_rate": 0.0018933345276398227, + "loss": 4.283267498016357, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.009412688203155994, + "learning_rate": 0.0018920454273232848, + "loss": 4.272198677062988, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.007799267303198576, + "learning_rate": 0.0018907560232280557, + "loss": 4.260125637054443, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.008747464045882225, + "learning_rate": 0.0018894663163929863, + "loss": 4.2539215087890625, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.008003434166312218, + "learning_rate": 0.0018881763078571689, + "loss": 4.245024681091309, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.007593846879899502, + "learning_rate": 0.0018868859986599415, + "loss": 4.237929344177246, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.008099420927464962, + "learning_rate": 0.0018855953898408815, + "loss": 4.2789106369018555, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.008270486257970333, + "learning_rate": 0.0018843044824398097, + "loss": 4.263522148132324, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.009260379709303379, + "learning_rate": 0.0018830132774967866, + "loss": 4.275835037231445, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.009289492852985859, + "learning_rate": 0.0018817217760521132, + "loss": 4.313778400421143, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.009493921883404255, + "learning_rate": 0.0018804299791463281, + "loss": 4.303672790527344, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.009020023047924042, + "learning_rate": 0.0018791378878202105, + "loss": 4.245473384857178, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.009446556679904461, + "learning_rate": 0.0018778455031147724, + "loss": 4.267327785491943, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.009499974548816681, + "learning_rate": 0.0018765528260712672, + "loss": 4.301205635070801, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.008399393409490585, + "learning_rate": 0.00187525985773118, + "loss": 4.307802677154541, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.009519701823592186, + "learning_rate": 0.0018739665991362327, + "loss": 4.284796714782715, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01125232595950365, + "learning_rate": 0.0018726730513283804, + "loss": 4.285930633544922, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.011608537286520004, + "learning_rate": 0.00187137921534981, + "loss": 4.257689476013184, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.01064533181488514, + "learning_rate": 0.001870085092242944, + "loss": 4.249063968658447, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.009854599833488464, + "learning_rate": 0.0018687906830504321, + "loss": 4.255159378051758, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.01121739111840725, + "learning_rate": 0.0018674959888151576, + "loss": 4.287777900695801, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.5675368309021, + "eval_runtime": 51.8555, + "eval_samples_per_second": 47.092, + "eval_steps_per_second": 1.485, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.0104464590549469, + "learning_rate": 0.0018662010105802314, + "loss": 4.206080436706543, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.012800181284546852, + "learning_rate": 0.0018649057493889945, + "loss": 4.213550567626953, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.013382901437580585, + "learning_rate": 0.001863610206285015, + "loss": 4.190858840942383, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.014347050338983536, + "learning_rate": 0.0018623143823120894, + "loss": 4.22884464263916, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.01347152329981327, + "learning_rate": 0.001861018278514239, + "loss": 4.215579032897949, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.012786377221345901, + "learning_rate": 0.0018597218959357112, + "loss": 4.174999237060547, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.014137202873826027, + "learning_rate": 0.001858425235620979, + "loss": 4.21709680557251, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.013864399865269661, + "learning_rate": 0.001857128298614737, + "loss": 4.2375030517578125, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.010397658683359623, + "learning_rate": 0.0018558310859619048, + "loss": 4.231938362121582, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.00888819433748722, + "learning_rate": 0.0018545335987076223, + "loss": 4.195562839508057, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.010698067955672741, + "learning_rate": 0.001853235837897252, + "loss": 4.22847318649292, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.00987589918076992, + "learning_rate": 0.0018519378045763766, + "loss": 4.2147040367126465, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.011451782658696175, + "learning_rate": 0.001850639499790798, + "loss": 4.179263114929199, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.011224704794585705, + "learning_rate": 0.0018493409245865368, + "loss": 4.23030948638916, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.00996861606836319, + "learning_rate": 0.0018480420800098313, + "loss": 4.194283962249756, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.012143119238317013, + "learning_rate": 0.0018467429671071374, + "loss": 4.221704959869385, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.013126783072948456, + "learning_rate": 0.0018454435869251268, + "loss": 4.187748908996582, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.013031953014433384, + "learning_rate": 0.001844143940510687, + "loss": 4.248250961303711, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.012868993915617466, + "learning_rate": 0.001842844028910919, + "loss": 4.20161247253418, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.010102041065692902, + "learning_rate": 0.0018415438531731395, + "loss": 4.213654518127441, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.009599820710718632, + "learning_rate": 0.0018402434143448747, + "loss": 4.206313133239746, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.011536866426467896, + "learning_rate": 0.0018389427134738658, + "loss": 4.237072944641113, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.009558089077472687, + "learning_rate": 0.0018376417516080636, + "loss": 4.182195663452148, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01056219544261694, + "learning_rate": 0.0018363405297956302, + "loss": 4.232585906982422, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.012673584744334221, + "learning_rate": 0.0018350390490849354, + "loss": 4.228830337524414, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.011215884238481522, + "learning_rate": 0.0018337373105245598, + "loss": 4.216056823730469, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.011281915940344334, + "learning_rate": 0.00183243531516329, + "loss": 4.233404159545898, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.010381227359175682, + "learning_rate": 0.0018311330640501204, + "loss": 4.182013034820557, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.00900198519229889, + "learning_rate": 0.0018298305582342513, + "loss": 4.215488433837891, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.010198601521551609, + "learning_rate": 0.0018285277987650876, + "loss": 4.19389533996582, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.009965402074158192, + "learning_rate": 0.0018272247866922401, + "loss": 4.218005657196045, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.009299935773015022, + "learning_rate": 0.0018259215230655202, + "loss": 4.223684310913086, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.008977951481938362, + "learning_rate": 0.0018246180089349462, + "loss": 4.215778350830078, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.009261307306587696, + "learning_rate": 0.0018233142453507344, + "loss": 4.215066909790039, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.008322088047862053, + "learning_rate": 0.0018220102333633044, + "loss": 4.2188005447387695, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.008716144599020481, + "learning_rate": 0.0018207059740232739, + "loss": 4.187867641448975, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.008897020481526852, + "learning_rate": 0.0018194014683814625, + "loss": 4.218034267425537, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.007298174314200878, + "learning_rate": 0.0018180967174888857, + "loss": 4.2017011642456055, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.008491789922118187, + "learning_rate": 0.0018167917223967582, + "loss": 4.239873886108398, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.008413378149271011, + "learning_rate": 0.0018154864841564908, + "loss": 4.234771251678467, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.008141471073031425, + "learning_rate": 0.0018141810038196909, + "loss": 4.185944080352783, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.009480944834649563, + "learning_rate": 0.0018128752824381603, + "loss": 4.206586837768555, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.0106047997251153, + "learning_rate": 0.001811569321063894, + "loss": 4.235287666320801, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.010663514956831932, + "learning_rate": 0.0018102631207490838, + "loss": 4.204899311065674, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.011473523452877998, + "learning_rate": 0.001808956682546109, + "loss": 4.253668308258057, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01165992021560669, + "learning_rate": 0.0018076500075075456, + "loss": 4.209845542907715, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.011692410334944725, + "learning_rate": 0.0018063430966861564, + "loss": 4.250368595123291, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01061506848782301, + "learning_rate": 0.0018050359511348965, + "loss": 4.193821907043457, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.010510941036045551, + "learning_rate": 0.0018037285719069097, + "loss": 4.254327774047852, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.011587041430175304, + "learning_rate": 0.001802420960055528, + "loss": 4.2838945388793945, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.01257420051842928, + "learning_rate": 0.0018011131166342693, + "loss": 4.258241653442383, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.00936233438551426, + "learning_rate": 0.0017998050426968407, + "loss": 4.201854705810547, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.010093127377331257, + "learning_rate": 0.0017984967392971338, + "loss": 4.193051338195801, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.010032009333372116, + "learning_rate": 0.0017971882074892236, + "loss": 4.230212211608887, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.009455587714910507, + "learning_rate": 0.0017958794483273726, + "loss": 4.2157673835754395, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.009549238719046116, + "learning_rate": 0.001794570462866023, + "loss": 4.260272026062012, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.011059699580073357, + "learning_rate": 0.0017932612521598013, + "loss": 4.214818000793457, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.011653627268970013, + "learning_rate": 0.0017919518172635146, + "loss": 4.205112457275391, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.012438314035534859, + "learning_rate": 0.001790642159232151, + "loss": 4.179045677185059, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.011234673671424389, + "learning_rate": 0.0017893322791208784, + "loss": 4.239317417144775, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.010923845693469048, + "learning_rate": 0.0017880221779850434, + "loss": 4.230571269989014, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.009219309315085411, + "learning_rate": 0.001786711856880171, + "loss": 4.229068756103516, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.009849398396909237, + "learning_rate": 0.0017854013168619638, + "loss": 4.254772186279297, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.01117763202637434, + "learning_rate": 0.001784090558986299, + "loss": 4.252647399902344, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.010664612054824829, + "learning_rate": 0.0017827795843092316, + "loss": 4.204596519470215, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.00847950391471386, + "learning_rate": 0.0017814683938869895, + "loss": 4.211061954498291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.007305726408958435, + "learning_rate": 0.0017801569887759753, + "loss": 4.181221008300781, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.007970666512846947, + "learning_rate": 0.0017788453700327653, + "loss": 4.237855434417725, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.008873614482581615, + "learning_rate": 0.0017775335387141054, + "loss": 4.229188919067383, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.009691286832094193, + "learning_rate": 0.0017762214958769156, + "loss": 4.205491542816162, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.009360672906041145, + "learning_rate": 0.0017749092425782852, + "loss": 4.233466625213623, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.009895884431898594, + "learning_rate": 0.0017735967798754723, + "loss": 4.187763214111328, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.535672187805176, + "eval_runtime": 53.1736, + "eval_samples_per_second": 45.925, + "eval_steps_per_second": 1.448, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.00966129545122385, + "learning_rate": 0.0017722841088259045, + "loss": 4.1451735496521, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.010733380913734436, + "learning_rate": 0.0017709712304871774, + "loss": 4.1616363525390625, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.0117328567430377, + "learning_rate": 0.0017696581459170535, + "loss": 4.126331806182861, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.012640298344194889, + "learning_rate": 0.0017683448561734607, + "loss": 4.163730621337891, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.013074076734483242, + "learning_rate": 0.0017670313623144934, + "loss": 4.199502944946289, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.01132719125598669, + "learning_rate": 0.0017657176653984097, + "loss": 4.143380165100098, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.010008333250880241, + "learning_rate": 0.0017644037664836319, + "loss": 4.170320510864258, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.011090578511357307, + "learning_rate": 0.0017630896666287438, + "loss": 4.1571478843688965, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.011047456413507462, + "learning_rate": 0.0017617753668924924, + "loss": 4.101892948150635, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.010098653845489025, + "learning_rate": 0.0017604608683337857, + "loss": 4.154077529907227, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.010499860160052776, + "learning_rate": 0.00175914617201169, + "loss": 4.116868019104004, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.010939318686723709, + "learning_rate": 0.0017578312789854338, + "loss": 4.159144878387451, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.010622999630868435, + "learning_rate": 0.0017565161903144023, + "loss": 4.171411991119385, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.0108700692653656, + "learning_rate": 0.0017552009070581386, + "loss": 4.100748062133789, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.009794988669455051, + "learning_rate": 0.0017538854302763423, + "loss": 4.185161590576172, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.007750354241579771, + "learning_rate": 0.0017525697610288702, + "loss": 4.149083137512207, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.009923134930431843, + "learning_rate": 0.001751253900375732, + "loss": 4.205166816711426, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.010811141692101955, + "learning_rate": 0.0017499378493770936, + "loss": 4.131865978240967, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.010868688113987446, + "learning_rate": 0.0017486216090932733, + "loss": 4.148791790008545, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.011224126443266869, + "learning_rate": 0.001747305180584743, + "loss": 4.152503490447998, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.013262554071843624, + "learning_rate": 0.001745988564912124, + "loss": 4.164926528930664, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.013089599087834358, + "learning_rate": 0.001744671763136191, + "loss": 4.1626458168029785, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.01327646616846323, + "learning_rate": 0.0017433547763178665, + "loss": 4.189356803894043, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.014539042487740517, + "learning_rate": 0.0017420376055182233, + "loss": 4.176387786865234, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.013269098475575447, + "learning_rate": 0.001740720251798482, + "loss": 4.180716514587402, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.010922391898930073, + "learning_rate": 0.0017394027162200105, + "loss": 4.163928031921387, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.013355271890759468, + "learning_rate": 0.0017380849998443243, + "loss": 4.150012016296387, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.014149201102554798, + "learning_rate": 0.0017367671037330828, + "loss": 4.173713207244873, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01354879979044199, + "learning_rate": 0.0017354490289480918, + "loss": 4.150563716888428, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.013788704760372639, + "learning_rate": 0.0017341307765512998, + "loss": 4.213858604431152, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.011597222648561, + "learning_rate": 0.0017328123476047994, + "loss": 4.166989326477051, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.01055675558745861, + "learning_rate": 0.001731493743170824, + "loss": 4.172226428985596, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.011600397527217865, + "learning_rate": 0.0017301749643117506, + "loss": 4.1656951904296875, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.01114651933312416, + "learning_rate": 0.0017288560120900943, + "loss": 4.189825057983398, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.009811583906412125, + "learning_rate": 0.0017275368875685126, + "loss": 4.181164741516113, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.009758968837559223, + "learning_rate": 0.0017262175918097992, + "loss": 4.163188934326172, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.00999707542359829, + "learning_rate": 0.001724898125876887, + "loss": 4.184117317199707, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.009528180584311485, + "learning_rate": 0.001723578490832846, + "loss": 4.16407585144043, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.009717358276247978, + "learning_rate": 0.0017222586877408826, + "loss": 4.209892272949219, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.010909050703048706, + "learning_rate": 0.0017209387176643377, + "loss": 4.19977331161499, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.011234750039875507, + "learning_rate": 0.0017196185816666875, + "loss": 4.19363260269165, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.010300721041858196, + "learning_rate": 0.0017182982808115425, + "loss": 4.142427921295166, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.00941646471619606, + "learning_rate": 0.001716977816162644, + "loss": 4.159372329711914, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.009304054081439972, + "learning_rate": 0.0017156571887838679, + "loss": 4.207922458648682, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.009496180340647697, + "learning_rate": 0.0017143363997392184, + "loss": 4.137894630432129, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.009257436729967594, + "learning_rate": 0.0017130154500928326, + "loss": 4.198753833770752, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.00900586973875761, + "learning_rate": 0.0017116943409089746, + "loss": 4.160684585571289, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.009398494847118855, + "learning_rate": 0.00171037307325204, + "loss": 4.1819353103637695, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.008633013814687729, + "learning_rate": 0.0017090516481865483, + "loss": 4.176833152770996, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.00898425467312336, + "learning_rate": 0.0017077300667771484, + "loss": 4.190657138824463, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.008182398043572903, + "learning_rate": 0.0017064083300886155, + "loss": 4.173429489135742, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.009589143097400665, + "learning_rate": 0.0017050864391858478, + "loss": 4.1559014320373535, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.010304383002221584, + "learning_rate": 0.0017037643951338698, + "loss": 4.155453681945801, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.009943748824298382, + "learning_rate": 0.0017024421989978275, + "loss": 4.119734287261963, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.009238033555448055, + "learning_rate": 0.001701119851842992, + "loss": 4.188974857330322, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.008058690465986729, + "learning_rate": 0.0016997973547347531, + "loss": 4.19934606552124, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.008485953323543072, + "learning_rate": 0.0016984747087386233, + "loss": 4.165615558624268, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.008476605638861656, + "learning_rate": 0.0016971519149202347, + "loss": 4.21021842956543, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.008588708005845547, + "learning_rate": 0.0016958289743453385, + "loss": 4.208064556121826, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.009223178029060364, + "learning_rate": 0.0016945058880798037, + "loss": 4.158413887023926, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.009196167811751366, + "learning_rate": 0.0016931826571896166, + "loss": 4.203187465667725, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.009488921612501144, + "learning_rate": 0.001691859282740882, + "loss": 4.178227424621582, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.009253161028027534, + "learning_rate": 0.0016905357657998167, + "loss": 4.131230354309082, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.010069048963487148, + "learning_rate": 0.0016892121074327559, + "loss": 4.210734844207764, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.009945095516741276, + "learning_rate": 0.001687888308706147, + "loss": 4.184042930603027, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.010483640246093273, + "learning_rate": 0.0016865643706865505, + "loss": 4.156919479370117, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.011560715734958649, + "learning_rate": 0.0016852402944406391, + "loss": 4.185520172119141, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.010001277551054955, + "learning_rate": 0.0016839160810351978, + "loss": 4.226828575134277, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.010335934348404408, + "learning_rate": 0.0016825917315371202, + "loss": 4.183531284332275, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.011322776786983013, + "learning_rate": 0.001681267247013412, + "loss": 4.179635047912598, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.009663194417953491, + "learning_rate": 0.001679942628531186, + "loss": 4.165654182434082, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.010590246878564358, + "learning_rate": 0.0016786178771576643, + "loss": 4.1539459228515625, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.511386394500732, + "eval_runtime": 51.7751, + "eval_samples_per_second": 47.166, + "eval_steps_per_second": 1.487, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.01254354789853096, + "learning_rate": 0.0016772929939601728, + "loss": 4.104002475738525, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.01300987508147955, + "learning_rate": 0.0016759679800061483, + "loss": 4.127395153045654, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.013325816951692104, + "learning_rate": 0.0016746428363631299, + "loss": 4.131229400634766, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.014302870258688927, + "learning_rate": 0.001673317564098761, + "loss": 4.079358100891113, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.013776068575680256, + "learning_rate": 0.0016719921642807905, + "loss": 4.132065773010254, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.010788560844957829, + "learning_rate": 0.0016706666379770689, + "loss": 4.134626388549805, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.009766630828380585, + "learning_rate": 0.0016693409862555493, + "loss": 4.1645426750183105, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.011760029010474682, + "learning_rate": 0.0016680152101842842, + "loss": 4.110664367675781, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.011047836393117905, + "learning_rate": 0.0016666893108314288, + "loss": 4.129222869873047, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.009989139623939991, + "learning_rate": 0.0016653632892652344, + "loss": 4.076835632324219, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.010033783502876759, + "learning_rate": 0.001664037146554054, + "loss": 4.11555290222168, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.00996427796781063, + "learning_rate": 0.001662710883766337, + "loss": 4.131463527679443, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.01044023409485817, + "learning_rate": 0.0016613845019706288, + "loss": 4.098750114440918, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.010033872909843922, + "learning_rate": 0.0016600580022355712, + "loss": 4.161447048187256, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.009329567663371563, + "learning_rate": 0.0016587313856299012, + "loss": 4.146491527557373, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.010061589069664478, + "learning_rate": 0.0016574046532224504, + "loss": 4.107517719268799, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.009491910226643085, + "learning_rate": 0.0016560778060821424, + "loss": 4.098226547241211, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.008529345504939556, + "learning_rate": 0.0016547508452779943, + "loss": 4.1247711181640625, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.00936342403292656, + "learning_rate": 0.0016534237718791136, + "loss": 4.120573043823242, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.010255641303956509, + "learning_rate": 0.0016520965869547008, + "loss": 4.089067459106445, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.011638758704066277, + "learning_rate": 0.001650769291574044, + "loss": 4.074847221374512, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.014707857742905617, + "learning_rate": 0.0016494418868065214, + "loss": 4.153377532958984, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.01650111749768257, + "learning_rate": 0.0016481143737215986, + "loss": 4.134062767028809, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.014833372086286545, + "learning_rate": 0.00164678675338883, + "loss": 4.1344828605651855, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.012257886119186878, + "learning_rate": 0.0016454590268778538, + "loss": 4.098813533782959, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.013508322648704052, + "learning_rate": 0.0016441311952583959, + "loss": 4.106297492980957, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.01256900280714035, + "learning_rate": 0.001642803259600267, + "loss": 4.143631935119629, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.009885406121611595, + "learning_rate": 0.00164147522097336, + "loss": 4.120885848999023, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.012004674412310123, + "learning_rate": 0.0016401470804476518, + "loss": 4.089243412017822, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01224028691649437, + "learning_rate": 0.0016388188390932012, + "loss": 4.077765464782715, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.012511290609836578, + "learning_rate": 0.001637490497980149, + "loss": 4.161346435546875, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.010701716877520084, + "learning_rate": 0.0016361620581787143, + "loss": 4.10698938369751, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.010713113471865654, + "learning_rate": 0.001634833520759198, + "loss": 4.114283561706543, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.011018907651305199, + "learning_rate": 0.001633504886791978, + "loss": 4.134284019470215, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.011279584839940071, + "learning_rate": 0.0016321761573475117, + "loss": 4.1456804275512695, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.010642931796610355, + "learning_rate": 0.001630847333496332, + "loss": 4.160918235778809, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.011069678701460361, + "learning_rate": 0.0016295184163090482, + "loss": 4.136838912963867, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.011007166467607021, + "learning_rate": 0.0016281894068563443, + "loss": 4.129153728485107, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.010055703110992908, + "learning_rate": 0.0016268603062089802, + "loss": 4.138665676116943, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.009839230217039585, + "learning_rate": 0.0016255311154377875, + "loss": 4.1366496086120605, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.011191432364284992, + "learning_rate": 0.0016242018356136714, + "loss": 4.129880905151367, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.012130341492593288, + "learning_rate": 0.0016228724678076095, + "loss": 4.1425981521606445, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.012707795947790146, + "learning_rate": 0.0016215430130906482, + "loss": 4.115452766418457, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.012707809917628765, + "learning_rate": 0.001620213472533906, + "loss": 4.122758865356445, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.01182397548109293, + "learning_rate": 0.001618883847208569, + "loss": 4.140275955200195, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.010865340940654278, + "learning_rate": 0.001617554138185893, + "loss": 4.137759208679199, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.010973592288792133, + "learning_rate": 0.0016162243465371998, + "loss": 4.138947010040283, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.011247226968407631, + "learning_rate": 0.001614894473333879, + "loss": 4.115516662597656, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.010837487876415253, + "learning_rate": 0.0016135645196473853, + "loss": 4.160427093505859, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.009096396155655384, + "learning_rate": 0.0016122344865492385, + "loss": 4.1200032234191895, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.010671247728168964, + "learning_rate": 0.0016109043751110223, + "loss": 4.097919464111328, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.0107562355697155, + "learning_rate": 0.0016095741864043831, + "loss": 4.096290588378906, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.008909327909350395, + "learning_rate": 0.0016082439215010305, + "loss": 4.171017646789551, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.010305124334990978, + "learning_rate": 0.0016069135814727343, + "loss": 4.1450276374816895, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.011235613375902176, + "learning_rate": 0.0016055831673913255, + "loss": 4.134704113006592, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.00825632456690073, + "learning_rate": 0.0016042526803286949, + "loss": 4.107443332672119, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.009056378155946732, + "learning_rate": 0.0016029221213567921, + "loss": 4.132656097412109, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.008835164830088615, + "learning_rate": 0.001601591491547624, + "loss": 4.124729633331299, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.00915333442389965, + "learning_rate": 0.0016002607919732557, + "loss": 4.119500160217285, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.008532539941370487, + "learning_rate": 0.0015989300237058066, + "loss": 4.141661643981934, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.009952311404049397, + "learning_rate": 0.0015975991878174533, + "loss": 4.140929222106934, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.008920297026634216, + "learning_rate": 0.0015962682853804268, + "loss": 4.155877113342285, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.008770049549639225, + "learning_rate": 0.0015949373174670104, + "loss": 4.126394271850586, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.0078084105625748634, + "learning_rate": 0.0015936062851495411, + "loss": 4.128876209259033, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.00816641841083765, + "learning_rate": 0.0015922751895004078, + "loss": 4.137410640716553, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.008895459584891796, + "learning_rate": 0.0015909440315920505, + "loss": 4.07957649230957, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.008551528677344322, + "learning_rate": 0.001589612812496958, + "loss": 4.104213714599609, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.008325488306581974, + "learning_rate": 0.0015882815332876707, + "loss": 4.1584792137146, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.009341313503682613, + "learning_rate": 0.0015869501950367744, + "loss": 4.1417436599731445, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.00999050959944725, + "learning_rate": 0.0015856187988169067, + "loss": 4.136049270629883, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.008889932185411453, + "learning_rate": 0.0015842873457007484, + "loss": 4.16257381439209, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.009109573438763618, + "learning_rate": 0.0015829558367610268, + "loss": 4.171441555023193, + "step": 1800 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 7.67403340333056e+17, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-1800/training_args.bin b/runs/baseline/checkpoint-1800/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-1800/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-2160/chat_template.jinja b/runs/baseline/checkpoint-2160/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-2160/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-2160/config.json b/runs/baseline/checkpoint-2160/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-2160/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-2160/generation_config.json b/runs/baseline/checkpoint-2160/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-2160/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-2160/model.safetensors b/runs/baseline/checkpoint-2160/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d6a9e9ce06875771d4d27d5d86b491e2f87bfe5a --- /dev/null +++ b/runs/baseline/checkpoint-2160/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b80a8ab78d106d60f21de0abe3f37e7784d6094041c106a808255e0e47720f82 +size 583356232 diff --git a/runs/baseline/checkpoint-2160/optimizer.pt b/runs/baseline/checkpoint-2160/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..8cfc40140a1ce500459fdb375065966569d80632 --- /dev/null +++ b/runs/baseline/checkpoint-2160/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f45e69b0c66324b605ce8625895e3bf9f80169892b256fb0c7418e13fee28c4d +size 1166825803 diff --git a/runs/baseline/checkpoint-2160/rng_state_0.pth b/runs/baseline/checkpoint-2160/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..86ca1aad69953afb19f9f8da4c01764c2f44d22b --- /dev/null +++ b/runs/baseline/checkpoint-2160/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:253e65ee29b5e996242fd0eef4ab59e40fe385a114b29c442de150cf74084b60 +size 14917 diff --git a/runs/baseline/checkpoint-2160/rng_state_1.pth b/runs/baseline/checkpoint-2160/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..3031af55ab1520df411513147e482ece04c4d4f6 --- /dev/null +++ b/runs/baseline/checkpoint-2160/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04c9d382d23a06579a44352e657618aaf7b647b0f5de65751c81300c8f96f42a +size 14917 diff --git a/runs/baseline/checkpoint-2160/scheduler.pt b/runs/baseline/checkpoint-2160/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..95e6d3366fe6af5dc063f7b8974daf6c734b2ea8 --- /dev/null +++ b/runs/baseline/checkpoint-2160/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63d15070a4477d2431b35b7fbfcc0f72990026ae265966082eb9b68e8bb7618c +size 1465 diff --git a/runs/baseline/checkpoint-2160/tokenizer.json b/runs/baseline/checkpoint-2160/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-2160/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-2160/tokenizer_config.json b/runs/baseline/checkpoint-2160/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-2160/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-2160/trainer_state.json b/runs/baseline/checkpoint-2160/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..ff194229364d154928e8b2e342c3a7a06a6e42ef --- /dev/null +++ b/runs/baseline/checkpoint-2160/trainer_state.json @@ -0,0 +1,15386 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 30.0, + "eval_steps": 500, + "global_step": 2160, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.641542434692383, + "eval_runtime": 51.6342, + "eval_samples_per_second": 47.294, + "eval_steps_per_second": 1.491, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.009694516658782959, + "learning_rate": 0.0020492498088767836, + "loss": 4.32808256149292, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.0097952950745821, + "learning_rate": 0.002048005909777114, + "loss": 4.280096054077148, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.010393359698355198, + "learning_rate": 0.0020467615812442106, + "loss": 4.317824363708496, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.00977372657507658, + "learning_rate": 0.0020455168242806054, + "loss": 4.3445329666137695, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.008568719029426575, + "learning_rate": 0.0020442716398891775, + "loss": 4.318026542663574, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.009732510894536972, + "learning_rate": 0.0020430260290731487, + "loss": 4.330609321594238, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.010429969057440758, + "learning_rate": 0.0020417799928360864, + "loss": 4.316794395446777, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.009292383678257465, + "learning_rate": 0.002040533532181898, + "loss": 4.317420959472656, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.011222455650568008, + "learning_rate": 0.0020392866481148356, + "loss": 4.292610168457031, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.013591725379228592, + "learning_rate": 0.0020380393416394894, + "loss": 4.325689315795898, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.013572026044130325, + "learning_rate": 0.002036791613760794, + "loss": 4.334960460662842, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.012986521236598492, + "learning_rate": 0.0020355434654840197, + "loss": 4.33829402923584, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.013321424834430218, + "learning_rate": 0.0020342948978147777, + "loss": 4.2926926612854, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.012307215481996536, + "learning_rate": 0.0020330459117590153, + "loss": 4.32757568359375, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.012909360229969025, + "learning_rate": 0.0020317965083230197, + "loss": 4.362889289855957, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017746610566973686, + "learning_rate": 0.0020305466885134115, + "loss": 4.3426313400268555, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.01696564070880413, + "learning_rate": 0.002029296453337148, + "loss": 4.3298516273498535, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.013348255306482315, + "learning_rate": 0.002028045803801522, + "loss": 4.360220432281494, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.013225145637989044, + "learning_rate": 0.002026794740914158, + "loss": 4.346250057220459, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.01569020375609398, + "learning_rate": 0.002025543265683016, + "loss": 4.308025360107422, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.018612321466207504, + "learning_rate": 0.002024291379116385, + "loss": 4.330190181732178, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.01619943603873253, + "learning_rate": 0.00202303908222289, + "loss": 4.339552402496338, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.015385213308036327, + "learning_rate": 0.0020217863760114826, + "loss": 4.3412184715271, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.018517112359404564, + "learning_rate": 0.002020533261491446, + "loss": 4.3289313316345215, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.014868509024381638, + "learning_rate": 0.0020192797396723903, + "loss": 4.359846115112305, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.014240248128771782, + "learning_rate": 0.0020180258115642578, + "loss": 4.369707107543945, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.018113158643245697, + "learning_rate": 0.0020167714781773143, + "loss": 4.362668514251709, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.01802460104227066, + "learning_rate": 0.002015516740522154, + "loss": 4.344742298126221, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.01414549071341753, + "learning_rate": 0.0020142615996096964, + "loss": 4.358080863952637, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.012157700955867767, + "learning_rate": 0.002013006056451186, + "loss": 4.347763538360596, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.011268534697592258, + "learning_rate": 0.0020117501120581905, + "loss": 4.335859298706055, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.010395628400146961, + "learning_rate": 0.002010493767442602, + "loss": 4.341609001159668, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.012147950939834118, + "learning_rate": 0.002009237023616635, + "loss": 4.328502655029297, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.00945813488215208, + "learning_rate": 0.0020079798815928252, + "loss": 4.311186790466309, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.009489500895142555, + "learning_rate": 0.0020067223423840294, + "loss": 4.371470928192139, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.009360029362142086, + "learning_rate": 0.0020054644070034234, + "loss": 4.33941125869751, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.010032779537141323, + "learning_rate": 0.0020042060764645037, + "loss": 4.365797519683838, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.009367046877741814, + "learning_rate": 0.002002947351781084, + "loss": 4.29443359375, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.008369733579456806, + "learning_rate": 0.0020016882339672967, + "loss": 4.294233322143555, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.010394785553216934, + "learning_rate": 0.0020004287240375893, + "loss": 4.352355003356934, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01068816613405943, + "learning_rate": 0.001999168823006728, + "loss": 4.342516899108887, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.009975973516702652, + "learning_rate": 0.0019979085318897905, + "loss": 4.346763610839844, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.009618151932954788, + "learning_rate": 0.001996647851702171, + "loss": 4.329115867614746, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.008399694226682186, + "learning_rate": 0.001995386783459577, + "loss": 4.28736686706543, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.008235268294811249, + "learning_rate": 0.001994125328178029, + "loss": 4.339872360229492, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.00836547277867794, + "learning_rate": 0.0019928634868738577, + "loss": 4.328156471252441, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.009619626216590405, + "learning_rate": 0.001991601260563706, + "loss": 4.337394714355469, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.009814038872718811, + "learning_rate": 0.0019903386502645275, + "loss": 4.325926780700684, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.009588954970240593, + "learning_rate": 0.001989075656993584, + "loss": 4.286477088928223, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.010471943765878677, + "learning_rate": 0.001987812281768447, + "loss": 4.322466850280762, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.010239035822451115, + "learning_rate": 0.001986548525606994, + "loss": 4.3487653732299805, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.008935568854212761, + "learning_rate": 0.001985284389527411, + "loss": 4.303601264953613, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.008624759502708912, + "learning_rate": 0.00198401987454819, + "loss": 4.34036922454834, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.009296547621488571, + "learning_rate": 0.0019827549816881277, + "loss": 4.273357391357422, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.009071138687431812, + "learning_rate": 0.0019814897119663248, + "loss": 4.321051597595215, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.008341389708220959, + "learning_rate": 0.0019802240664021876, + "loss": 4.343242645263672, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.007881715893745422, + "learning_rate": 0.0019789580460154227, + "loss": 4.32067346572876, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.008750553242862225, + "learning_rate": 0.0019776916518260404, + "loss": 4.330263614654541, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.008591460064053535, + "learning_rate": 0.001976424884854351, + "loss": 4.311957359313965, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.008084543980658054, + "learning_rate": 0.001975157746120966, + "loss": 4.331319808959961, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.008036594837903976, + "learning_rate": 0.001973890236646797, + "loss": 4.382656097412109, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.00805772002786398, + "learning_rate": 0.0019726223574530527, + "loss": 4.297247886657715, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.009019420482218266, + "learning_rate": 0.0019713541095612405, + "loss": 4.326526165008545, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.009953883476555347, + "learning_rate": 0.0019700854939931653, + "loss": 4.327356338500977, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.009340492077171803, + "learning_rate": 0.0019688165117709273, + "loss": 4.343993186950684, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.010395721532404423, + "learning_rate": 0.0019675471639169226, + "loss": 4.284934997558594, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.009946230798959732, + "learning_rate": 0.0019662774514538415, + "loss": 4.288936614990234, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.010981100611388683, + "learning_rate": 0.0019650073754046695, + "loss": 4.339935779571533, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.009688783437013626, + "learning_rate": 0.0019637369367926827, + "loss": 4.381425857543945, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.009442255832254887, + "learning_rate": 0.001962466136641452, + "loss": 4.304185390472412, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.009371291846036911, + "learning_rate": 0.0019611949759748364, + "loss": 4.306120872497559, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.009544814005494118, + "learning_rate": 0.001959923455816988, + "loss": 4.326251983642578, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.602745532989502, + "eval_runtime": 53.5041, + "eval_samples_per_second": 45.641, + "eval_steps_per_second": 1.439, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.01010364294052124, + "learning_rate": 0.0019586515771923474, + "loss": 4.24491024017334, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.012722729705274105, + "learning_rate": 0.0019573793411256446, + "loss": 4.257739543914795, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.01307204831391573, + "learning_rate": 0.0019561067486418967, + "loss": 4.291684150695801, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.011215966194868088, + "learning_rate": 0.0019548338007664098, + "loss": 4.237947463989258, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.012943019159138203, + "learning_rate": 0.0019535604985247727, + "loss": 4.242859363555908, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.01337282732129097, + "learning_rate": 0.0019522868429428636, + "loss": 4.238934516906738, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.013697991147637367, + "learning_rate": 0.0019510128350468438, + "loss": 4.285760402679443, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.012182159349322319, + "learning_rate": 0.001949738475863158, + "loss": 4.290044784545898, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.012620339170098305, + "learning_rate": 0.0019484637664185346, + "loss": 4.267187118530273, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.011284958571195602, + "learning_rate": 0.0019471887077399837, + "loss": 4.26662540435791, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.01161450706422329, + "learning_rate": 0.001945913300854797, + "loss": 4.283192157745361, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.010022968985140324, + "learning_rate": 0.001944637546790548, + "loss": 4.295843124389648, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.010341878980398178, + "learning_rate": 0.001943361446575087, + "loss": 4.295636177062988, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.011549185030162334, + "learning_rate": 0.0019420850012365457, + "loss": 4.288768768310547, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.012518810108304024, + "learning_rate": 0.0019408082118033331, + "loss": 4.229338645935059, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.012035328894853592, + "learning_rate": 0.0019395310793041353, + "loss": 4.258639335632324, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.011193782091140747, + "learning_rate": 0.0019382536047679146, + "loss": 4.239264488220215, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.009930136613547802, + "learning_rate": 0.0019369757892239096, + "loss": 4.252286434173584, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.00953229982405901, + "learning_rate": 0.0019356976337016327, + "loss": 4.293435096740723, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.010618340224027634, + "learning_rate": 0.0019344191392308716, + "loss": 4.235588550567627, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.008806651458144188, + "learning_rate": 0.0019331403068416856, + "loss": 4.26716423034668, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.008934796787798405, + "learning_rate": 0.0019318611375644065, + "loss": 4.258462905883789, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.009283283725380898, + "learning_rate": 0.0019305816324296386, + "loss": 4.257868766784668, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.009224812500178814, + "learning_rate": 0.0019293017924682555, + "loss": 4.258486747741699, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.008957969956099987, + "learning_rate": 0.0019280216187114016, + "loss": 4.309563636779785, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.010284051299095154, + "learning_rate": 0.0019267411121904894, + "loss": 4.275271415710449, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.009167762473225594, + "learning_rate": 0.0019254602739372001, + "loss": 4.238452911376953, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.009411636739969254, + "learning_rate": 0.0019241791049834816, + "loss": 4.2735595703125, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.009044171310961246, + "learning_rate": 0.0019228976063615497, + "loss": 4.27017879486084, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.008840068243443966, + "learning_rate": 0.0019216157791038834, + "loss": 4.28773307800293, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.009512359276413918, + "learning_rate": 0.001920333624243228, + "loss": 4.266008377075195, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.009946667589247227, + "learning_rate": 0.001919051142812593, + "loss": 4.279062271118164, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.00899894442409277, + "learning_rate": 0.0019177683358452505, + "loss": 4.277679443359375, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.00909668579697609, + "learning_rate": 0.0019164852043747338, + "loss": 4.311397552490234, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.009787327609956264, + "learning_rate": 0.0019152017494348408, + "loss": 4.30875825881958, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.008983899839222431, + "learning_rate": 0.001913917972059626, + "loss": 4.259649276733398, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.008411582559347153, + "learning_rate": 0.0019126338732834071, + "loss": 4.259811878204346, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.009162918664515018, + "learning_rate": 0.0019113494541407579, + "loss": 4.271515846252441, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.011304561980068684, + "learning_rate": 0.0019100647156665125, + "loss": 4.269217491149902, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.012483394704759121, + "learning_rate": 0.0019087796588957617, + "loss": 4.2949299812316895, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.012091380544006824, + "learning_rate": 0.0019074942848638522, + "loss": 4.289752006530762, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.010011567734181881, + "learning_rate": 0.0019062085946063869, + "loss": 4.294095993041992, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.008766217157244682, + "learning_rate": 0.0019049225891592226, + "loss": 4.256425857543945, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.012862488627433777, + "learning_rate": 0.001903636269558471, + "loss": 4.287847995758057, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.013917063362896442, + "learning_rate": 0.0019023496368404972, + "loss": 4.282784938812256, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.012129023671150208, + "learning_rate": 0.0019010626920419173, + "loss": 4.2486419677734375, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.010348426178097725, + "learning_rate": 0.0018997754361995986, + "loss": 4.289836883544922, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.011821693740785122, + "learning_rate": 0.001898487870350662, + "loss": 4.214344024658203, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.013136724010109901, + "learning_rate": 0.0018971999955324742, + "loss": 4.24208402633667, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.013173168525099754, + "learning_rate": 0.0018959118127826538, + "loss": 4.270164489746094, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.011018159799277782, + "learning_rate": 0.0018946233231390656, + "loss": 4.2571916580200195, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.01018658559769392, + "learning_rate": 0.0018933345276398227, + "loss": 4.283267498016357, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.009412688203155994, + "learning_rate": 0.0018920454273232848, + "loss": 4.272198677062988, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.007799267303198576, + "learning_rate": 0.0018907560232280557, + "loss": 4.260125637054443, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.008747464045882225, + "learning_rate": 0.0018894663163929863, + "loss": 4.2539215087890625, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.008003434166312218, + "learning_rate": 0.0018881763078571689, + "loss": 4.245024681091309, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.007593846879899502, + "learning_rate": 0.0018868859986599415, + "loss": 4.237929344177246, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.008099420927464962, + "learning_rate": 0.0018855953898408815, + "loss": 4.2789106369018555, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.008270486257970333, + "learning_rate": 0.0018843044824398097, + "loss": 4.263522148132324, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.009260379709303379, + "learning_rate": 0.0018830132774967866, + "loss": 4.275835037231445, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.009289492852985859, + "learning_rate": 0.0018817217760521132, + "loss": 4.313778400421143, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.009493921883404255, + "learning_rate": 0.0018804299791463281, + "loss": 4.303672790527344, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.009020023047924042, + "learning_rate": 0.0018791378878202105, + "loss": 4.245473384857178, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.009446556679904461, + "learning_rate": 0.0018778455031147724, + "loss": 4.267327785491943, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.009499974548816681, + "learning_rate": 0.0018765528260712672, + "loss": 4.301205635070801, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.008399393409490585, + "learning_rate": 0.00187525985773118, + "loss": 4.307802677154541, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.009519701823592186, + "learning_rate": 0.0018739665991362327, + "loss": 4.284796714782715, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01125232595950365, + "learning_rate": 0.0018726730513283804, + "loss": 4.285930633544922, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.011608537286520004, + "learning_rate": 0.00187137921534981, + "loss": 4.257689476013184, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.01064533181488514, + "learning_rate": 0.001870085092242944, + "loss": 4.249063968658447, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.009854599833488464, + "learning_rate": 0.0018687906830504321, + "loss": 4.255159378051758, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.01121739111840725, + "learning_rate": 0.0018674959888151576, + "loss": 4.287777900695801, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.5675368309021, + "eval_runtime": 51.8555, + "eval_samples_per_second": 47.092, + "eval_steps_per_second": 1.485, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.0104464590549469, + "learning_rate": 0.0018662010105802314, + "loss": 4.206080436706543, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.012800181284546852, + "learning_rate": 0.0018649057493889945, + "loss": 4.213550567626953, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.013382901437580585, + "learning_rate": 0.001863610206285015, + "loss": 4.190858840942383, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.014347050338983536, + "learning_rate": 0.0018623143823120894, + "loss": 4.22884464263916, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.01347152329981327, + "learning_rate": 0.001861018278514239, + "loss": 4.215579032897949, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.012786377221345901, + "learning_rate": 0.0018597218959357112, + "loss": 4.174999237060547, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.014137202873826027, + "learning_rate": 0.001858425235620979, + "loss": 4.21709680557251, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.013864399865269661, + "learning_rate": 0.001857128298614737, + "loss": 4.2375030517578125, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.010397658683359623, + "learning_rate": 0.0018558310859619048, + "loss": 4.231938362121582, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.00888819433748722, + "learning_rate": 0.0018545335987076223, + "loss": 4.195562839508057, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.010698067955672741, + "learning_rate": 0.001853235837897252, + "loss": 4.22847318649292, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.00987589918076992, + "learning_rate": 0.0018519378045763766, + "loss": 4.2147040367126465, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.011451782658696175, + "learning_rate": 0.001850639499790798, + "loss": 4.179263114929199, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.011224704794585705, + "learning_rate": 0.0018493409245865368, + "loss": 4.23030948638916, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.00996861606836319, + "learning_rate": 0.0018480420800098313, + "loss": 4.194283962249756, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.012143119238317013, + "learning_rate": 0.0018467429671071374, + "loss": 4.221704959869385, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.013126783072948456, + "learning_rate": 0.0018454435869251268, + "loss": 4.187748908996582, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.013031953014433384, + "learning_rate": 0.001844143940510687, + "loss": 4.248250961303711, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.012868993915617466, + "learning_rate": 0.001842844028910919, + "loss": 4.20161247253418, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.010102041065692902, + "learning_rate": 0.0018415438531731395, + "loss": 4.213654518127441, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.009599820710718632, + "learning_rate": 0.0018402434143448747, + "loss": 4.206313133239746, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.011536866426467896, + "learning_rate": 0.0018389427134738658, + "loss": 4.237072944641113, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.009558089077472687, + "learning_rate": 0.0018376417516080636, + "loss": 4.182195663452148, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01056219544261694, + "learning_rate": 0.0018363405297956302, + "loss": 4.232585906982422, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.012673584744334221, + "learning_rate": 0.0018350390490849354, + "loss": 4.228830337524414, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.011215884238481522, + "learning_rate": 0.0018337373105245598, + "loss": 4.216056823730469, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.011281915940344334, + "learning_rate": 0.00183243531516329, + "loss": 4.233404159545898, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.010381227359175682, + "learning_rate": 0.0018311330640501204, + "loss": 4.182013034820557, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.00900198519229889, + "learning_rate": 0.0018298305582342513, + "loss": 4.215488433837891, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.010198601521551609, + "learning_rate": 0.0018285277987650876, + "loss": 4.19389533996582, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.009965402074158192, + "learning_rate": 0.0018272247866922401, + "loss": 4.218005657196045, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.009299935773015022, + "learning_rate": 0.0018259215230655202, + "loss": 4.223684310913086, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.008977951481938362, + "learning_rate": 0.0018246180089349462, + "loss": 4.215778350830078, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.009261307306587696, + "learning_rate": 0.0018233142453507344, + "loss": 4.215066909790039, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.008322088047862053, + "learning_rate": 0.0018220102333633044, + "loss": 4.2188005447387695, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.008716144599020481, + "learning_rate": 0.0018207059740232739, + "loss": 4.187867641448975, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.008897020481526852, + "learning_rate": 0.0018194014683814625, + "loss": 4.218034267425537, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.007298174314200878, + "learning_rate": 0.0018180967174888857, + "loss": 4.2017011642456055, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.008491789922118187, + "learning_rate": 0.0018167917223967582, + "loss": 4.239873886108398, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.008413378149271011, + "learning_rate": 0.0018154864841564908, + "loss": 4.234771251678467, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.008141471073031425, + "learning_rate": 0.0018141810038196909, + "loss": 4.185944080352783, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.009480944834649563, + "learning_rate": 0.0018128752824381603, + "loss": 4.206586837768555, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.0106047997251153, + "learning_rate": 0.001811569321063894, + "loss": 4.235287666320801, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.010663514956831932, + "learning_rate": 0.0018102631207490838, + "loss": 4.204899311065674, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.011473523452877998, + "learning_rate": 0.001808956682546109, + "loss": 4.253668308258057, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01165992021560669, + "learning_rate": 0.0018076500075075456, + "loss": 4.209845542907715, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.011692410334944725, + "learning_rate": 0.0018063430966861564, + "loss": 4.250368595123291, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01061506848782301, + "learning_rate": 0.0018050359511348965, + "loss": 4.193821907043457, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.010510941036045551, + "learning_rate": 0.0018037285719069097, + "loss": 4.254327774047852, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.011587041430175304, + "learning_rate": 0.001802420960055528, + "loss": 4.2838945388793945, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.01257420051842928, + "learning_rate": 0.0018011131166342693, + "loss": 4.258241653442383, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.00936233438551426, + "learning_rate": 0.0017998050426968407, + "loss": 4.201854705810547, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.010093127377331257, + "learning_rate": 0.0017984967392971338, + "loss": 4.193051338195801, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.010032009333372116, + "learning_rate": 0.0017971882074892236, + "loss": 4.230212211608887, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.009455587714910507, + "learning_rate": 0.0017958794483273726, + "loss": 4.2157673835754395, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.009549238719046116, + "learning_rate": 0.001794570462866023, + "loss": 4.260272026062012, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.011059699580073357, + "learning_rate": 0.0017932612521598013, + "loss": 4.214818000793457, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.011653627268970013, + "learning_rate": 0.0017919518172635146, + "loss": 4.205112457275391, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.012438314035534859, + "learning_rate": 0.001790642159232151, + "loss": 4.179045677185059, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.011234673671424389, + "learning_rate": 0.0017893322791208784, + "loss": 4.239317417144775, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.010923845693469048, + "learning_rate": 0.0017880221779850434, + "loss": 4.230571269989014, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.009219309315085411, + "learning_rate": 0.001786711856880171, + "loss": 4.229068756103516, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.009849398396909237, + "learning_rate": 0.0017854013168619638, + "loss": 4.254772186279297, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.01117763202637434, + "learning_rate": 0.001784090558986299, + "loss": 4.252647399902344, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.010664612054824829, + "learning_rate": 0.0017827795843092316, + "loss": 4.204596519470215, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.00847950391471386, + "learning_rate": 0.0017814683938869895, + "loss": 4.211061954498291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.007305726408958435, + "learning_rate": 0.0017801569887759753, + "loss": 4.181221008300781, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.007970666512846947, + "learning_rate": 0.0017788453700327653, + "loss": 4.237855434417725, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.008873614482581615, + "learning_rate": 0.0017775335387141054, + "loss": 4.229188919067383, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.009691286832094193, + "learning_rate": 0.0017762214958769156, + "loss": 4.205491542816162, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.009360672906041145, + "learning_rate": 0.0017749092425782852, + "loss": 4.233466625213623, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.009895884431898594, + "learning_rate": 0.0017735967798754723, + "loss": 4.187763214111328, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.535672187805176, + "eval_runtime": 53.1736, + "eval_samples_per_second": 45.925, + "eval_steps_per_second": 1.448, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.00966129545122385, + "learning_rate": 0.0017722841088259045, + "loss": 4.1451735496521, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.010733380913734436, + "learning_rate": 0.0017709712304871774, + "loss": 4.1616363525390625, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.0117328567430377, + "learning_rate": 0.0017696581459170535, + "loss": 4.126331806182861, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.012640298344194889, + "learning_rate": 0.0017683448561734607, + "loss": 4.163730621337891, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.013074076734483242, + "learning_rate": 0.0017670313623144934, + "loss": 4.199502944946289, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.01132719125598669, + "learning_rate": 0.0017657176653984097, + "loss": 4.143380165100098, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.010008333250880241, + "learning_rate": 0.0017644037664836319, + "loss": 4.170320510864258, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.011090578511357307, + "learning_rate": 0.0017630896666287438, + "loss": 4.1571478843688965, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.011047456413507462, + "learning_rate": 0.0017617753668924924, + "loss": 4.101892948150635, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.010098653845489025, + "learning_rate": 0.0017604608683337857, + "loss": 4.154077529907227, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.010499860160052776, + "learning_rate": 0.00175914617201169, + "loss": 4.116868019104004, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.010939318686723709, + "learning_rate": 0.0017578312789854338, + "loss": 4.159144878387451, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.010622999630868435, + "learning_rate": 0.0017565161903144023, + "loss": 4.171411991119385, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.0108700692653656, + "learning_rate": 0.0017552009070581386, + "loss": 4.100748062133789, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.009794988669455051, + "learning_rate": 0.0017538854302763423, + "loss": 4.185161590576172, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.007750354241579771, + "learning_rate": 0.0017525697610288702, + "loss": 4.149083137512207, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.009923134930431843, + "learning_rate": 0.001751253900375732, + "loss": 4.205166816711426, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.010811141692101955, + "learning_rate": 0.0017499378493770936, + "loss": 4.131865978240967, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.010868688113987446, + "learning_rate": 0.0017486216090932733, + "loss": 4.148791790008545, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.011224126443266869, + "learning_rate": 0.001747305180584743, + "loss": 4.152503490447998, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.013262554071843624, + "learning_rate": 0.001745988564912124, + "loss": 4.164926528930664, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.013089599087834358, + "learning_rate": 0.001744671763136191, + "loss": 4.1626458168029785, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.01327646616846323, + "learning_rate": 0.0017433547763178665, + "loss": 4.189356803894043, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.014539042487740517, + "learning_rate": 0.0017420376055182233, + "loss": 4.176387786865234, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.013269098475575447, + "learning_rate": 0.001740720251798482, + "loss": 4.180716514587402, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.010922391898930073, + "learning_rate": 0.0017394027162200105, + "loss": 4.163928031921387, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.013355271890759468, + "learning_rate": 0.0017380849998443243, + "loss": 4.150012016296387, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.014149201102554798, + "learning_rate": 0.0017367671037330828, + "loss": 4.173713207244873, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01354879979044199, + "learning_rate": 0.0017354490289480918, + "loss": 4.150563716888428, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.013788704760372639, + "learning_rate": 0.0017341307765512998, + "loss": 4.213858604431152, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.011597222648561, + "learning_rate": 0.0017328123476047994, + "loss": 4.166989326477051, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.01055675558745861, + "learning_rate": 0.001731493743170824, + "loss": 4.172226428985596, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.011600397527217865, + "learning_rate": 0.0017301749643117506, + "loss": 4.1656951904296875, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.01114651933312416, + "learning_rate": 0.0017288560120900943, + "loss": 4.189825057983398, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.009811583906412125, + "learning_rate": 0.0017275368875685126, + "loss": 4.181164741516113, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.009758968837559223, + "learning_rate": 0.0017262175918097992, + "loss": 4.163188934326172, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.00999707542359829, + "learning_rate": 0.001724898125876887, + "loss": 4.184117317199707, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.009528180584311485, + "learning_rate": 0.001723578490832846, + "loss": 4.16407585144043, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.009717358276247978, + "learning_rate": 0.0017222586877408826, + "loss": 4.209892272949219, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.010909050703048706, + "learning_rate": 0.0017209387176643377, + "loss": 4.19977331161499, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.011234750039875507, + "learning_rate": 0.0017196185816666875, + "loss": 4.19363260269165, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.010300721041858196, + "learning_rate": 0.0017182982808115425, + "loss": 4.142427921295166, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.00941646471619606, + "learning_rate": 0.001716977816162644, + "loss": 4.159372329711914, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.009304054081439972, + "learning_rate": 0.0017156571887838679, + "loss": 4.207922458648682, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.009496180340647697, + "learning_rate": 0.0017143363997392184, + "loss": 4.137894630432129, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.009257436729967594, + "learning_rate": 0.0017130154500928326, + "loss": 4.198753833770752, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.00900586973875761, + "learning_rate": 0.0017116943409089746, + "loss": 4.160684585571289, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.009398494847118855, + "learning_rate": 0.00171037307325204, + "loss": 4.1819353103637695, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.008633013814687729, + "learning_rate": 0.0017090516481865483, + "loss": 4.176833152770996, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.00898425467312336, + "learning_rate": 0.0017077300667771484, + "loss": 4.190657138824463, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.008182398043572903, + "learning_rate": 0.0017064083300886155, + "loss": 4.173429489135742, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.009589143097400665, + "learning_rate": 0.0017050864391858478, + "loss": 4.1559014320373535, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.010304383002221584, + "learning_rate": 0.0017037643951338698, + "loss": 4.155453681945801, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.009943748824298382, + "learning_rate": 0.0017024421989978275, + "loss": 4.119734287261963, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.009238033555448055, + "learning_rate": 0.001701119851842992, + "loss": 4.188974857330322, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.008058690465986729, + "learning_rate": 0.0016997973547347531, + "loss": 4.19934606552124, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.008485953323543072, + "learning_rate": 0.0016984747087386233, + "loss": 4.165615558624268, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.008476605638861656, + "learning_rate": 0.0016971519149202347, + "loss": 4.21021842956543, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.008588708005845547, + "learning_rate": 0.0016958289743453385, + "loss": 4.208064556121826, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.009223178029060364, + "learning_rate": 0.0016945058880798037, + "loss": 4.158413887023926, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.009196167811751366, + "learning_rate": 0.0016931826571896166, + "loss": 4.203187465667725, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.009488921612501144, + "learning_rate": 0.001691859282740882, + "loss": 4.178227424621582, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.009253161028027534, + "learning_rate": 0.0016905357657998167, + "loss": 4.131230354309082, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.010069048963487148, + "learning_rate": 0.0016892121074327559, + "loss": 4.210734844207764, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.009945095516741276, + "learning_rate": 0.001687888308706147, + "loss": 4.184042930603027, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.010483640246093273, + "learning_rate": 0.0016865643706865505, + "loss": 4.156919479370117, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.011560715734958649, + "learning_rate": 0.0016852402944406391, + "loss": 4.185520172119141, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.010001277551054955, + "learning_rate": 0.0016839160810351978, + "loss": 4.226828575134277, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.010335934348404408, + "learning_rate": 0.0016825917315371202, + "loss": 4.183531284332275, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.011322776786983013, + "learning_rate": 0.001681267247013412, + "loss": 4.179635047912598, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.009663194417953491, + "learning_rate": 0.001679942628531186, + "loss": 4.165654182434082, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.010590246878564358, + "learning_rate": 0.0016786178771576643, + "loss": 4.1539459228515625, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.511386394500732, + "eval_runtime": 51.7751, + "eval_samples_per_second": 47.166, + "eval_steps_per_second": 1.487, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.01254354789853096, + "learning_rate": 0.0016772929939601728, + "loss": 4.104002475738525, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.01300987508147955, + "learning_rate": 0.0016759679800061483, + "loss": 4.127395153045654, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.013325816951692104, + "learning_rate": 0.0016746428363631299, + "loss": 4.131229400634766, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.014302870258688927, + "learning_rate": 0.001673317564098761, + "loss": 4.079358100891113, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.013776068575680256, + "learning_rate": 0.0016719921642807905, + "loss": 4.132065773010254, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.010788560844957829, + "learning_rate": 0.0016706666379770689, + "loss": 4.134626388549805, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.009766630828380585, + "learning_rate": 0.0016693409862555493, + "loss": 4.1645426750183105, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.011760029010474682, + "learning_rate": 0.0016680152101842842, + "loss": 4.110664367675781, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.011047836393117905, + "learning_rate": 0.0016666893108314288, + "loss": 4.129222869873047, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.009989139623939991, + "learning_rate": 0.0016653632892652344, + "loss": 4.076835632324219, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.010033783502876759, + "learning_rate": 0.001664037146554054, + "loss": 4.11555290222168, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.00996427796781063, + "learning_rate": 0.001662710883766337, + "loss": 4.131463527679443, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.01044023409485817, + "learning_rate": 0.0016613845019706288, + "loss": 4.098750114440918, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.010033872909843922, + "learning_rate": 0.0016600580022355712, + "loss": 4.161447048187256, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.009329567663371563, + "learning_rate": 0.0016587313856299012, + "loss": 4.146491527557373, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.010061589069664478, + "learning_rate": 0.0016574046532224504, + "loss": 4.107517719268799, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.009491910226643085, + "learning_rate": 0.0016560778060821424, + "loss": 4.098226547241211, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.008529345504939556, + "learning_rate": 0.0016547508452779943, + "loss": 4.1247711181640625, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.00936342403292656, + "learning_rate": 0.0016534237718791136, + "loss": 4.120573043823242, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.010255641303956509, + "learning_rate": 0.0016520965869547008, + "loss": 4.089067459106445, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.011638758704066277, + "learning_rate": 0.001650769291574044, + "loss": 4.074847221374512, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.014707857742905617, + "learning_rate": 0.0016494418868065214, + "loss": 4.153377532958984, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.01650111749768257, + "learning_rate": 0.0016481143737215986, + "loss": 4.134062767028809, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.014833372086286545, + "learning_rate": 0.00164678675338883, + "loss": 4.1344828605651855, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.012257886119186878, + "learning_rate": 0.0016454590268778538, + "loss": 4.098813533782959, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.013508322648704052, + "learning_rate": 0.0016441311952583959, + "loss": 4.106297492980957, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.01256900280714035, + "learning_rate": 0.001642803259600267, + "loss": 4.143631935119629, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.009885406121611595, + "learning_rate": 0.00164147522097336, + "loss": 4.120885848999023, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.012004674412310123, + "learning_rate": 0.0016401470804476518, + "loss": 4.089243412017822, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01224028691649437, + "learning_rate": 0.0016388188390932012, + "loss": 4.077765464782715, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.012511290609836578, + "learning_rate": 0.001637490497980149, + "loss": 4.161346435546875, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.010701716877520084, + "learning_rate": 0.0016361620581787143, + "loss": 4.10698938369751, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.010713113471865654, + "learning_rate": 0.001634833520759198, + "loss": 4.114283561706543, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.011018907651305199, + "learning_rate": 0.001633504886791978, + "loss": 4.134284019470215, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.011279584839940071, + "learning_rate": 0.0016321761573475117, + "loss": 4.1456804275512695, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.010642931796610355, + "learning_rate": 0.001630847333496332, + "loss": 4.160918235778809, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.011069678701460361, + "learning_rate": 0.0016295184163090482, + "loss": 4.136838912963867, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.011007166467607021, + "learning_rate": 0.0016281894068563443, + "loss": 4.129153728485107, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.010055703110992908, + "learning_rate": 0.0016268603062089802, + "loss": 4.138665676116943, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.009839230217039585, + "learning_rate": 0.0016255311154377875, + "loss": 4.1366496086120605, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.011191432364284992, + "learning_rate": 0.0016242018356136714, + "loss": 4.129880905151367, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.012130341492593288, + "learning_rate": 0.0016228724678076095, + "loss": 4.1425981521606445, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.012707795947790146, + "learning_rate": 0.0016215430130906482, + "loss": 4.115452766418457, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.012707809917628765, + "learning_rate": 0.001620213472533906, + "loss": 4.122758865356445, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.01182397548109293, + "learning_rate": 0.001618883847208569, + "loss": 4.140275955200195, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.010865340940654278, + "learning_rate": 0.001617554138185893, + "loss": 4.137759208679199, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.010973592288792133, + "learning_rate": 0.0016162243465371998, + "loss": 4.138947010040283, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.011247226968407631, + "learning_rate": 0.001614894473333879, + "loss": 4.115516662597656, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.010837487876415253, + "learning_rate": 0.0016135645196473853, + "loss": 4.160427093505859, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.009096396155655384, + "learning_rate": 0.0016122344865492385, + "loss": 4.1200032234191895, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.010671247728168964, + "learning_rate": 0.0016109043751110223, + "loss": 4.097919464111328, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.0107562355697155, + "learning_rate": 0.0016095741864043831, + "loss": 4.096290588378906, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.008909327909350395, + "learning_rate": 0.0016082439215010305, + "loss": 4.171017646789551, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.010305124334990978, + "learning_rate": 0.0016069135814727343, + "loss": 4.1450276374816895, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.011235613375902176, + "learning_rate": 0.0016055831673913255, + "loss": 4.134704113006592, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.00825632456690073, + "learning_rate": 0.0016042526803286949, + "loss": 4.107443332672119, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.009056378155946732, + "learning_rate": 0.0016029221213567921, + "loss": 4.132656097412109, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.008835164830088615, + "learning_rate": 0.001601591491547624, + "loss": 4.124729633331299, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.00915333442389965, + "learning_rate": 0.0016002607919732557, + "loss": 4.119500160217285, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.008532539941370487, + "learning_rate": 0.0015989300237058066, + "loss": 4.141661643981934, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.009952311404049397, + "learning_rate": 0.0015975991878174533, + "loss": 4.140929222106934, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.008920297026634216, + "learning_rate": 0.0015962682853804268, + "loss": 4.155877113342285, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.008770049549639225, + "learning_rate": 0.0015949373174670104, + "loss": 4.126394271850586, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.0078084105625748634, + "learning_rate": 0.0015936062851495411, + "loss": 4.128876209259033, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.00816641841083765, + "learning_rate": 0.0015922751895004078, + "loss": 4.137410640716553, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.008895459584891796, + "learning_rate": 0.0015909440315920505, + "loss": 4.07957649230957, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.008551528677344322, + "learning_rate": 0.001589612812496958, + "loss": 4.104213714599609, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.008325488306581974, + "learning_rate": 0.0015882815332876707, + "loss": 4.1584792137146, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.009341313503682613, + "learning_rate": 0.0015869501950367744, + "loss": 4.1417436599731445, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.00999050959944725, + "learning_rate": 0.0015856187988169067, + "loss": 4.136049270629883, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.008889932185411453, + "learning_rate": 0.0015842873457007484, + "loss": 4.16257381439209, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.009109573438763618, + "learning_rate": 0.0015829558367610268, + "loss": 4.171441555023193, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.485942840576172, + "eval_runtime": 52.1051, + "eval_samples_per_second": 46.867, + "eval_steps_per_second": 1.478, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.009881424717605114, + "learning_rate": 0.001581624273070515, + "loss": 4.059608459472656, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.011156721040606499, + "learning_rate": 0.0015802926557020294, + "loss": 4.047014236450195, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.012326992116868496, + "learning_rate": 0.0015789609857284309, + "loss": 4.06541109085083, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.01332208700478077, + "learning_rate": 0.0015776292642226213, + "loss": 4.042714595794678, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.012720128521323204, + "learning_rate": 0.0015762974922575446, + "loss": 4.098389625549316, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.010723215527832508, + "learning_rate": 0.0015749656709061854, + "loss": 4.059185981750488, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.010605281218886375, + "learning_rate": 0.001573633801241569, + "loss": 4.099538326263428, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.01101530622690916, + "learning_rate": 0.0015723018843367576, + "loss": 4.095674514770508, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.01135838869959116, + "learning_rate": 0.0015709699212648532, + "loss": 4.070535659790039, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.012119271792471409, + "learning_rate": 0.0015696379130989936, + "loss": 4.060605049133301, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.012783961370587349, + "learning_rate": 0.001568305860912355, + "loss": 4.075931072235107, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.011486220173537731, + "learning_rate": 0.0015669737657781461, + "loss": 4.075855255126953, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.01037552859634161, + "learning_rate": 0.0015656416287696136, + "loss": 4.087729454040527, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.009802743792533875, + "learning_rate": 0.0015643094509600354, + "loss": 4.0855607986450195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.010804438032209873, + "learning_rate": 0.001562977233422723, + "loss": 4.064528465270996, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.009505728259682655, + "learning_rate": 0.0015616449772310206, + "loss": 4.036775588989258, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.010634485632181168, + "learning_rate": 0.0015603126834583024, + "loss": 4.069923400878906, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.011392096057534218, + "learning_rate": 0.0015589803531779736, + "loss": 4.0812177658081055, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.011851722374558449, + "learning_rate": 0.0015576479874634683, + "loss": 4.09183406829834, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.011205444112420082, + "learning_rate": 0.0015563155873882502, + "loss": 4.07825231552124, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.009676349349319935, + "learning_rate": 0.0015549831540258093, + "loss": 4.052826404571533, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.009376165457069874, + "learning_rate": 0.0015536506884496638, + "loss": 4.02100133895874, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.00936955213546753, + "learning_rate": 0.0015523181917333564, + "loss": 4.049526214599609, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.00937784556299448, + "learning_rate": 0.0015509856649504564, + "loss": 4.096356391906738, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.010014008730649948, + "learning_rate": 0.0015496531091745555, + "loss": 4.03254508972168, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.009909634478390217, + "learning_rate": 0.0015483205254792705, + "loss": 4.067630767822266, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.009319501928985119, + "learning_rate": 0.0015469879149382399, + "loss": 4.094167709350586, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.008634721860289574, + "learning_rate": 0.0015456552786251244, + "loss": 4.122404098510742, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.009412454441189766, + "learning_rate": 0.0015443226176136049, + "loss": 4.081546783447266, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.010102108120918274, + "learning_rate": 0.0015429899329773814, + "loss": 4.049839019775391, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.010921654291450977, + "learning_rate": 0.0015416572257901746, + "loss": 4.097097396850586, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.010607713833451271, + "learning_rate": 0.0015403244971257226, + "loss": 4.084481239318848, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.011887960135936737, + "learning_rate": 0.0015389917480577806, + "loss": 4.074968338012695, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01089888159185648, + "learning_rate": 0.0015376589796601203, + "loss": 4.062132835388184, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.009934183210134506, + "learning_rate": 0.0015363261930065298, + "loss": 4.128092288970947, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.00878534372895956, + "learning_rate": 0.0015349933891708105, + "loss": 4.069051742553711, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.009928155690431595, + "learning_rate": 0.0015336605692267788, + "loss": 4.073912620544434, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.011091411113739014, + "learning_rate": 0.0015323277342482637, + "loss": 4.125944137573242, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.010314172133803368, + "learning_rate": 0.0015309948853091055, + "loss": 4.062620162963867, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.010423967614769936, + "learning_rate": 0.001529662023483158, + "loss": 4.101737976074219, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.009822997264564037, + "learning_rate": 0.0015283291498442817, + "loss": 4.064520359039307, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.009757095016539097, + "learning_rate": 0.001526996265466351, + "loss": 4.100624084472656, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.008650688454508781, + "learning_rate": 0.001525663371423246, + "loss": 4.109685897827148, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.007378234062343836, + "learning_rate": 0.0015243304687888553, + "loss": 4.044726848602295, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.008194336667656898, + "learning_rate": 0.001522997558637074, + "loss": 4.097923278808594, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.009764892049133778, + "learning_rate": 0.0015216646420418048, + "loss": 4.102148056030273, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.009242868050932884, + "learning_rate": 0.001520331720076953, + "loss": 4.068563461303711, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.010409324429929256, + "learning_rate": 0.001518998793816431, + "loss": 4.065944194793701, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.009215258061885834, + "learning_rate": 0.0015176658643341527, + "loss": 4.10223388671875, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.008065393194556236, + "learning_rate": 0.0015163329327040366, + "loss": 4.099432945251465, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.008498288691043854, + "learning_rate": 0.001515, + "loss": 4.109854698181152, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.008919150568544865, + "learning_rate": 0.0015136670672959633, + "loss": 4.128247261047363, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.009104429744184017, + "learning_rate": 0.0015123341356658474, + "loss": 4.068622589111328, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.008558365516364574, + "learning_rate": 0.0015110012061835692, + "loss": 4.114214897155762, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.007865076884627342, + "learning_rate": 0.0015096682799230477, + "loss": 4.11155366897583, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.007848534733057022, + "learning_rate": 0.0015083353579581958, + "loss": 4.045694351196289, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.008238030597567558, + "learning_rate": 0.0015070024413629264, + "loss": 4.093930721282959, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.007536398712545633, + "learning_rate": 0.0015056695312111448, + "loss": 4.106235504150391, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.008156892843544483, + "learning_rate": 0.001504336628576754, + "loss": 4.13060998916626, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.008144439198076725, + "learning_rate": 0.0015030037345336487, + "loss": 4.091304779052734, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.007757358253002167, + "learning_rate": 0.001501670850155718, + "loss": 4.069276809692383, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.00860623549669981, + "learning_rate": 0.0015003379765168425, + "loss": 4.107658386230469, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.008614457212388515, + "learning_rate": 0.0014990051146908949, + "loss": 4.106705665588379, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.009120592847466469, + "learning_rate": 0.001497672265751737, + "loss": 4.0891923904418945, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.009446980431675911, + "learning_rate": 0.0014963394307732213, + "loss": 4.098847389221191, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.010494766756892204, + "learning_rate": 0.0014950066108291894, + "loss": 4.115325450897217, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.010057658888399601, + "learning_rate": 0.0014936738069934706, + "loss": 4.099984169006348, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.009071771055459976, + "learning_rate": 0.0014923410203398796, + "loss": 4.091543674468994, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.009937075898051262, + "learning_rate": 0.00149100825194222, + "loss": 4.086492538452148, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.010222986340522766, + "learning_rate": 0.0014896755028742778, + "loss": 4.063841819763184, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.010513762012124062, + "learning_rate": 0.001488342774209826, + "loss": 4.11991548538208, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.009530222974717617, + "learning_rate": 0.0014870100670226192, + "loss": 4.043334007263184, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.467459678649902, + "eval_runtime": 54.1628, + "eval_samples_per_second": 45.086, + "eval_steps_per_second": 1.422, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.010274187661707401, + "learning_rate": 0.0014856773823863955, + "loss": 4.049561500549316, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.011448442004621029, + "learning_rate": 0.001484344721374876, + "loss": 4.019484996795654, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.011361192911863327, + "learning_rate": 0.0014830120850617599, + "loss": 4.0429887771606445, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.010250156745314598, + "learning_rate": 0.0014816794745207299, + "loss": 4.012682914733887, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.010024400427937508, + "learning_rate": 0.0014803468908254449, + "loss": 3.979409694671631, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.01153535582125187, + "learning_rate": 0.0014790143350495446, + "loss": 4.045780181884766, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.012632817961275578, + "learning_rate": 0.001477681808266644, + "loss": 4.057641983032227, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.013758293353021145, + "learning_rate": 0.0014763493115503366, + "loss": 4.061607360839844, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.014041583053767681, + "learning_rate": 0.001475016845974191, + "loss": 4.035029411315918, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.013629456982016563, + "learning_rate": 0.0014736844126117497, + "loss": 4.062169075012207, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.011023213155567646, + "learning_rate": 0.0014723520125365323, + "loss": 4.02731990814209, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.01001750584691763, + "learning_rate": 0.0014710196468220266, + "loss": 4.019882678985596, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.008321646600961685, + "learning_rate": 0.001469687316541698, + "loss": 4.005446434020996, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.008935946971178055, + "learning_rate": 0.0014683550227689796, + "loss": 4.017102241516113, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.009137130342423916, + "learning_rate": 0.001467022766577277, + "loss": 4.024102210998535, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.00805022194981575, + "learning_rate": 0.001465690549039965, + "loss": 3.9935250282287598, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.009204058907926083, + "learning_rate": 0.0014643583712303865, + "loss": 4.042818546295166, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.009812815114855766, + "learning_rate": 0.001463026234221854, + "loss": 4.0252485275268555, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.009955782443284988, + "learning_rate": 0.0014616941390876454, + "loss": 3.9839768409729004, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.009374645538628101, + "learning_rate": 0.001460362086901007, + "loss": 4.040245056152344, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.010008739307522774, + "learning_rate": 0.0014590300787351471, + "loss": 4.0306596755981445, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.009014200419187546, + "learning_rate": 0.001457698115663243, + "loss": 4.028713226318359, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.008381523191928864, + "learning_rate": 0.0014563661987584313, + "loss": 4.024718284606934, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.008627948351204395, + "learning_rate": 0.0014550343290938143, + "loss": 4.02303409576416, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.010257705114781857, + "learning_rate": 0.0014537025077424556, + "loss": 4.061058044433594, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.010759209282696247, + "learning_rate": 0.0014523707357773788, + "loss": 4.064718246459961, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.009966139681637287, + "learning_rate": 0.0014510390142715697, + "loss": 3.9919447898864746, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.00997584406286478, + "learning_rate": 0.001449707344297971, + "loss": 4.0431599617004395, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.010846944525837898, + "learning_rate": 0.0014483757269294858, + "loss": 4.0222697257995605, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.00991352554410696, + "learning_rate": 0.0014470441632389735, + "loss": 4.059438705444336, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.008994397707283497, + "learning_rate": 0.0014457126542992518, + "loss": 4.021494388580322, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.009567267261445522, + "learning_rate": 0.0014443812011830934, + "loss": 4.070810317993164, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.009199011139571667, + "learning_rate": 0.0014430498049632255, + "loss": 4.020938873291016, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.010440350510179996, + "learning_rate": 0.0014417184667123303, + "loss": 4.062529563903809, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.009962482377886772, + "learning_rate": 0.0014403871875030424, + "loss": 4.0687150955200195, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.009389379993081093, + "learning_rate": 0.0014390559684079502, + "loss": 4.055185794830322, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.011469941586256027, + "learning_rate": 0.0014377248104995924, + "loss": 4.041023254394531, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.011034650728106499, + "learning_rate": 0.0014363937148504588, + "loss": 4.054254531860352, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.010651151649653912, + "learning_rate": 0.0014350626825329897, + "loss": 4.037753582000732, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.010747930034995079, + "learning_rate": 0.0014337317146195733, + "loss": 4.059938907623291, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.00987576600164175, + "learning_rate": 0.0014324008121825468, + "loss": 4.075366973876953, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.008504371158778667, + "learning_rate": 0.0014310699762941936, + "loss": 4.026755332946777, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.009445052593946457, + "learning_rate": 0.001429739208026745, + "loss": 4.034431457519531, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.010560023598372936, + "learning_rate": 0.0014284085084523763, + "loss": 4.101158618927002, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.009920580312609673, + "learning_rate": 0.0014270778786432078, + "loss": 4.041321754455566, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.009854268282651901, + "learning_rate": 0.0014257473196713053, + "loss": 4.049882411956787, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.009825135581195354, + "learning_rate": 0.0014244168326086745, + "loss": 4.067440986633301, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.010506784543395042, + "learning_rate": 0.0014230864185272663, + "loss": 4.042476654052734, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.009527767077088356, + "learning_rate": 0.0014217560784989696, + "loss": 4.055248260498047, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.01013826485723257, + "learning_rate": 0.001420425813595617, + "loss": 4.05246639251709, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.011020111851394176, + "learning_rate": 0.0014190956248889778, + "loss": 4.032987117767334, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.010647117160260677, + "learning_rate": 0.0014177655134507614, + "loss": 4.093589782714844, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.009484780952334404, + "learning_rate": 0.001416435480352615, + "loss": 4.017437934875488, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.0114199910312891, + "learning_rate": 0.0014151055266661211, + "loss": 4.080643653869629, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.011137664318084717, + "learning_rate": 0.0014137756534628007, + "loss": 4.069380760192871, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.008972682990133762, + "learning_rate": 0.0014124458618141073, + "loss": 4.051383018493652, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.010311643593013287, + "learning_rate": 0.0014111161527914317, + "loss": 4.048953533172607, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.01113453321158886, + "learning_rate": 0.0014097865274660944, + "loss": 4.063544273376465, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.01091703586280346, + "learning_rate": 0.0014084569869093518, + "loss": 4.006721496582031, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.010672449134290218, + "learning_rate": 0.0014071275321923906, + "loss": 4.020220756530762, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.009682527743279934, + "learning_rate": 0.001405798164386328, + "loss": 4.057307243347168, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.009808676317334175, + "learning_rate": 0.0014044688845622126, + "loss": 4.105412483215332, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.010351189412176609, + "learning_rate": 0.0014031396937910202, + "loss": 4.066454887390137, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.010038275271654129, + "learning_rate": 0.001401810593143656, + "loss": 4.056752681732178, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01016094908118248, + "learning_rate": 0.001400481583690952, + "loss": 4.097933292388916, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.008683547377586365, + "learning_rate": 0.0013991526665036679, + "loss": 4.0386152267456055, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.009556332603096962, + "learning_rate": 0.0013978238426524882, + "loss": 4.05835485458374, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.009597044438123703, + "learning_rate": 0.001396495113208022, + "loss": 4.052363872528076, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.010021139867603779, + "learning_rate": 0.0013951664792408026, + "loss": 4.083431243896484, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.010611358098685741, + "learning_rate": 0.0013938379418212863, + "loss": 4.042330265045166, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.009132112376391888, + "learning_rate": 0.001392509502019852, + "loss": 4.058825969696045, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.009808709844946861, + "learning_rate": 0.0013911811609067992, + "loss": 4.093010902404785, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.4473161697387695, + "eval_runtime": 51.7738, + "eval_samples_per_second": 47.167, + "eval_steps_per_second": 1.487, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.011859052814543247, + "learning_rate": 0.0013898529195523488, + "loss": 3.9572651386260986, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.013210238888859749, + "learning_rate": 0.0013885247790266406, + "loss": 3.9812350273132324, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.012981324456632137, + "learning_rate": 0.001387196740399733, + "loss": 3.987725019454956, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.01424382347613573, + "learning_rate": 0.001385868804741604, + "loss": 3.9719889163970947, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.0154642965644598, + "learning_rate": 0.0013845409731221465, + "loss": 3.984248161315918, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.014779271557927132, + "learning_rate": 0.0013832132466111706, + "loss": 4.002912521362305, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.012921188026666641, + "learning_rate": 0.0013818856262784013, + "loss": 3.9588499069213867, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01165325753390789, + "learning_rate": 0.0013805581131934792, + "loss": 3.993288993835449, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.008908954448997974, + "learning_rate": 0.0013792307084259563, + "loss": 3.987151861190796, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.012040507979691029, + "learning_rate": 0.0013779034130452991, + "loss": 3.9675703048706055, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.013464349322021008, + "learning_rate": 0.0013765762281208868, + "loss": 3.9901437759399414, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.009626250714063644, + "learning_rate": 0.0013752491547220063, + "loss": 3.9373281002044678, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.00897257961332798, + "learning_rate": 0.0013739221939178582, + "loss": 4.00804328918457, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.010689414106309414, + "learning_rate": 0.00137259534677755, + "loss": 4.020804405212402, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.010462156496942043, + "learning_rate": 0.0013712686143700988, + "loss": 4.028973579406738, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.009135857224464417, + "learning_rate": 0.001369941997764429, + "loss": 3.9798216819763184, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.009298899210989475, + "learning_rate": 0.0013686154980293713, + "loss": 4.018845081329346, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.00979928020387888, + "learning_rate": 0.0013672891162336636, + "loss": 3.9946775436401367, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.008175542578101158, + "learning_rate": 0.0013659628534459457, + "loss": 3.981858491897583, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.008302642032504082, + "learning_rate": 0.0013646367107347662, + "loss": 3.991128444671631, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.008516009896993637, + "learning_rate": 0.0013633106891685718, + "loss": 4.004990100860596, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.007936482317745686, + "learning_rate": 0.0013619847898157164, + "loss": 3.995734214782715, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.008317600004374981, + "learning_rate": 0.001360659013744451, + "loss": 4.040653228759766, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.008481282740831375, + "learning_rate": 0.001359333362022931, + "loss": 3.969860553741455, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.008944948203861713, + "learning_rate": 0.0013580078357192096, + "loss": 3.9908909797668457, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.008897850289940834, + "learning_rate": 0.001356682435901239, + "loss": 4.04786491394043, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.009527456946671009, + "learning_rate": 0.001355357163636871, + "loss": 4.017423629760742, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.011714582331478596, + "learning_rate": 0.001354032019993852, + "loss": 3.9990692138671875, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.011083807796239853, + "learning_rate": 0.0013527070060398274, + "loss": 3.995265483856201, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.009878288954496384, + "learning_rate": 0.0013513821228423363, + "loss": 4.0364532470703125, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.009727329947054386, + "learning_rate": 0.0013500573714688136, + "loss": 4.035953521728516, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.009482033550739288, + "learning_rate": 0.0013487327529865879, + "loss": 4.015855312347412, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.009037062525749207, + "learning_rate": 0.0013474082684628798, + "loss": 4.024164199829102, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.009094092063605785, + "learning_rate": 0.0013460839189648028, + "loss": 4.037765979766846, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.008230654522776604, + "learning_rate": 0.0013447597055593612, + "loss": 4.006317615509033, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.008491289801895618, + "learning_rate": 0.0013434356293134501, + "loss": 4.042603969573975, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.008907693438231945, + "learning_rate": 0.0013421116912938533, + "loss": 4.008098602294922, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.008897818624973297, + "learning_rate": 0.0013407878925672438, + "loss": 3.9892945289611816, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.00957555789500475, + "learning_rate": 0.0013394642342001832, + "loss": 4.046009063720703, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.009948568418622017, + "learning_rate": 0.0013381407172591185, + "loss": 3.9760642051696777, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.01014955248683691, + "learning_rate": 0.0013368173428103835, + "loss": 4.023215293884277, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.008578895591199398, + "learning_rate": 0.0013354941119201966, + "loss": 4.002819061279297, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.008729875087738037, + "learning_rate": 0.001334171025654662, + "loss": 4.0340657234191895, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.009556280449032784, + "learning_rate": 0.0013328480850797655, + "loss": 3.983076333999634, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.009733820334076881, + "learning_rate": 0.0013315252912613766, + "loss": 4.066578388214111, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.009025252424180508, + "learning_rate": 0.0013302026452652474, + "loss": 3.9980697631835938, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.008485320024192333, + "learning_rate": 0.001328880148157008, + "loss": 4.018826961517334, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.009539696387946606, + "learning_rate": 0.0013275578010021726, + "loss": 4.011737823486328, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.010881016030907631, + "learning_rate": 0.0013262356048661305, + "loss": 4.007291793823242, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.010362236760556698, + "learning_rate": 0.0013249135608141523, + "loss": 3.9929394721984863, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.010296608321368694, + "learning_rate": 0.0013235916699113849, + "loss": 4.0015130043029785, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.011211174540221691, + "learning_rate": 0.0013222699332228517, + "loss": 4.005324363708496, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.009743994101881981, + "learning_rate": 0.0013209483518134525, + "loss": 4.0278143882751465, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.008563914336264133, + "learning_rate": 0.0013196269267479607, + "loss": 3.9997801780700684, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.009078902192413807, + "learning_rate": 0.0013183056590910255, + "loss": 4.026200294494629, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.009618895128369331, + "learning_rate": 0.0013169845499071675, + "loss": 4.027050018310547, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.008920644409954548, + "learning_rate": 0.0013156636002607822, + "loss": 4.000490665435791, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.008849202655255795, + "learning_rate": 0.0013143428112161323, + "loss": 4.044985294342041, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.009303621016442776, + "learning_rate": 0.001313022183837356, + "loss": 4.046281814575195, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.01076227892190218, + "learning_rate": 0.0013117017191884579, + "loss": 4.030576705932617, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.010400841943919659, + "learning_rate": 0.0013103814183333127, + "loss": 3.980081558227539, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.009459166787564754, + "learning_rate": 0.001309061282335663, + "loss": 4.014824867248535, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.009054969064891338, + "learning_rate": 0.001307741312259118, + "loss": 4.0485944747924805, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.009003991261124611, + "learning_rate": 0.0013064215091671545, + "loss": 3.9990320205688477, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.008959448896348476, + "learning_rate": 0.001305101874123113, + "loss": 4.038620948791504, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.009102243930101395, + "learning_rate": 0.0013037824081902015, + "loss": 4.030906677246094, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.008599337190389633, + "learning_rate": 0.0013024631124314877, + "loss": 4.02648401260376, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.008771977387368679, + "learning_rate": 0.0013011439879099056, + "loss": 4.027151107788086, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.009082595817744732, + "learning_rate": 0.00129982503568825, + "loss": 4.0194478034973145, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.009025354869663715, + "learning_rate": 0.0012985062568291763, + "loss": 4.044382095336914, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.00875102635473013, + "learning_rate": 0.0012971876523952014, + "loss": 4.0257568359375, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.009312704205513, + "learning_rate": 0.0012958692234487006, + "loss": 3.9964494705200195, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.431707859039307, + "eval_runtime": 50.7311, + "eval_samples_per_second": 48.136, + "eval_steps_per_second": 1.518, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.00992550328373909, + "learning_rate": 0.0012945509710519088, + "loss": 3.9253792762756348, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.009920453652739525, + "learning_rate": 0.0012932328962669173, + "loss": 3.9438819885253906, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.009437218308448792, + "learning_rate": 0.0012919150001556756, + "loss": 3.960963249206543, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.009616893716156483, + "learning_rate": 0.0012905972837799896, + "loss": 3.969301223754883, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.009107022546231747, + "learning_rate": 0.0012892797482015185, + "loss": 3.9626078605651855, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.011317167431116104, + "learning_rate": 0.0012879623944817772, + "loss": 3.946483850479126, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.010536770336329937, + "learning_rate": 0.001286645223682134, + "loss": 3.959566116333008, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.009349033236503601, + "learning_rate": 0.0012853282368638097, + "loss": 3.964435338973999, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.009950080886483192, + "learning_rate": 0.0012840114350878762, + "loss": 3.9729413986206055, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.010833023115992546, + "learning_rate": 0.001282694819415257, + "loss": 4.006421089172363, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.011538525111973286, + "learning_rate": 0.0012813783909067266, + "loss": 3.9558355808258057, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.010671315714716911, + "learning_rate": 0.0012800621506229065, + "loss": 4.0052337646484375, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.010835555382072926, + "learning_rate": 0.0012787460996242686, + "loss": 3.958641767501831, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.01082367543131113, + "learning_rate": 0.0012774302389711304, + "loss": 3.9818902015686035, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.009194346144795418, + "learning_rate": 0.0012761145697236583, + "loss": 3.9688854217529297, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.007859935984015465, + "learning_rate": 0.001274799092941862, + "loss": 3.9730639457702637, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.008534743450582027, + "learning_rate": 0.0012734838096855978, + "loss": 3.975003719329834, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.009369445964694023, + "learning_rate": 0.0012721687210145666, + "loss": 3.972907304763794, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.0090147340670228, + "learning_rate": 0.0012708538279883098, + "loss": 3.964256525039673, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.007745139300823212, + "learning_rate": 0.0012695391316662151, + "loss": 3.9504833221435547, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.008920101448893547, + "learning_rate": 0.0012682246331075075, + "loss": 3.9735097885131836, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.008987226523458958, + "learning_rate": 0.0012669103333712568, + "loss": 3.9482927322387695, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.00841289572417736, + "learning_rate": 0.0012655962335163685, + "loss": 3.965221881866455, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.008016116917133331, + "learning_rate": 0.0012642823346015902, + "loss": 3.9390206336975098, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.008470152504742146, + "learning_rate": 0.0012629686376855068, + "loss": 3.9257445335388184, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.008637234568595886, + "learning_rate": 0.0012616551438265394, + "loss": 3.9999356269836426, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.008036923594772816, + "learning_rate": 0.0012603418540829473, + "loss": 3.9755172729492188, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.008908843621611595, + "learning_rate": 0.0012590287695128232, + "loss": 3.9248545169830322, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.008743193000555038, + "learning_rate": 0.001257715891174096, + "loss": 3.9619390964508057, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.00834199320524931, + "learning_rate": 0.001256403220124528, + "loss": 3.961569309234619, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.009148893877863884, + "learning_rate": 0.001255090757421715, + "loss": 3.972104549407959, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.00925084762275219, + "learning_rate": 0.0012537785041230843, + "loss": 4.000775337219238, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.008659154176712036, + "learning_rate": 0.0012524664612858947, + "loss": 3.954564332962036, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.007787859998643398, + "learning_rate": 0.0012511546299672355, + "loss": 3.9498860836029053, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.00817918125540018, + "learning_rate": 0.0012498430112240246, + "loss": 4.003897666931152, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.009282548911869526, + "learning_rate": 0.001248531606113011, + "loss": 3.9621925354003906, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.009375610388815403, + "learning_rate": 0.001247220415690769, + "loss": 3.9977524280548096, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.009650378488004208, + "learning_rate": 0.001245909441013701, + "loss": 3.9882829189300537, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.009350919164717197, + "learning_rate": 0.0012445986831380366, + "loss": 4.019596099853516, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.008726303465664387, + "learning_rate": 0.0012432881431198289, + "loss": 3.9933621883392334, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.00829484686255455, + "learning_rate": 0.0012419778220149568, + "loss": 4.001237869262695, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.008744452148675919, + "learning_rate": 0.0012406677208791218, + "loss": 3.9323887825012207, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.008170492947101593, + "learning_rate": 0.0012393578407678496, + "loss": 3.925581455230713, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.007843422703444958, + "learning_rate": 0.001238048182736486, + "loss": 4.001023769378662, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.008914309553802013, + "learning_rate": 0.0012367387478401987, + "loss": 3.9804537296295166, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.009736751206219196, + "learning_rate": 0.0012354295371339774, + "loss": 3.9831767082214355, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.010061381384730339, + "learning_rate": 0.0012341205516726273, + "loss": 3.9462435245513916, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.010350692085921764, + "learning_rate": 0.0012328117925107765, + "loss": 4.012863636016846, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01039902213960886, + "learning_rate": 0.0012315032607028667, + "loss": 3.9811301231384277, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.011423371732234955, + "learning_rate": 0.0012301949573031595, + "loss": 3.9773542881011963, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.012449614703655243, + "learning_rate": 0.001228886883365731, + "loss": 3.969600200653076, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.010361429303884506, + "learning_rate": 0.0012275790399444726, + "loss": 4.035360336303711, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.009603082202374935, + "learning_rate": 0.0012262714280930904, + "loss": 3.9840846061706543, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.009183845482766628, + "learning_rate": 0.0012249640488651036, + "loss": 4.031542778015137, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.009694446809589863, + "learning_rate": 0.0012236569033138441, + "loss": 3.9935057163238525, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.0097816726192832, + "learning_rate": 0.0012223499924924545, + "loss": 3.9851176738739014, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.009413184598088264, + "learning_rate": 0.001221043317453891, + "loss": 3.9642205238342285, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.009960711002349854, + "learning_rate": 0.0012197368792509166, + "loss": 3.9373817443847656, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.010124670341610909, + "learning_rate": 0.0012184306789361059, + "loss": 3.958115577697754, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.011087656021118164, + "learning_rate": 0.00121712471756184, + "loss": 3.9607057571411133, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.009812800213694572, + "learning_rate": 0.001215818996180309, + "loss": 4.00722599029541, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.010305573232471943, + "learning_rate": 0.0012145135158435094, + "loss": 3.98728084564209, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.01019787136465311, + "learning_rate": 0.0012132082776032422, + "loss": 3.981172800064087, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.009260181337594986, + "learning_rate": 0.001211903282511115, + "loss": 4.034339904785156, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.00972998607903719, + "learning_rate": 0.001210598531618538, + "loss": 4.022885322570801, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.009853586554527283, + "learning_rate": 0.0012092940259767267, + "loss": 4.000235080718994, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.00961449183523655, + "learning_rate": 0.0012079897666366962, + "loss": 3.9951353073120117, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.010257176123559475, + "learning_rate": 0.0012066857546492657, + "loss": 3.9801371097564697, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.009443561546504498, + "learning_rate": 0.001205381991065054, + "loss": 3.9596502780914307, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.010389018803834915, + "learning_rate": 0.0012040784769344795, + "loss": 3.980527639389038, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.010322567075490952, + "learning_rate": 0.0012027752133077607, + "loss": 3.996258497238159, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.009987728670239449, + "learning_rate": 0.0012014722012349125, + "loss": 3.9493203163146973, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.421086311340332, + "eval_runtime": 52.5623, + "eval_samples_per_second": 46.459, + "eval_steps_per_second": 1.465, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.009405115619301796, + "learning_rate": 0.0012001694417657493, + "loss": 3.927506923675537, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.009974643588066101, + "learning_rate": 0.00119886693594988, + "loss": 3.9373550415039062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.011331096291542053, + "learning_rate": 0.0011975646848367098, + "loss": 3.935594081878662, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.010538448579609394, + "learning_rate": 0.0011962626894754405, + "loss": 3.911579132080078, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.00989572424441576, + "learning_rate": 0.0011949609509150645, + "loss": 3.92525577545166, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.010533696971833706, + "learning_rate": 0.0011936594702043704, + "loss": 3.898075580596924, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.01051455456763506, + "learning_rate": 0.0011923582483919365, + "loss": 3.93463134765625, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.009662075899541378, + "learning_rate": 0.0011910572865261348, + "loss": 3.898221731185913, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.008874096907675266, + "learning_rate": 0.0011897565856551257, + "loss": 3.926787853240967, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.008126807399094105, + "learning_rate": 0.0011884561468268606, + "loss": 3.9158413410186768, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.007915553636848927, + "learning_rate": 0.001187155971089081, + "loss": 3.8891496658325195, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.008312922902405262, + "learning_rate": 0.0011858560594893129, + "loss": 3.932219982147217, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.008892328478395939, + "learning_rate": 0.0011845564130748734, + "loss": 3.8866336345672607, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.009187526069581509, + "learning_rate": 0.0011832570328928628, + "loss": 3.9464707374572754, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.008615494705736637, + "learning_rate": 0.0011819579199901693, + "loss": 3.9396629333496094, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.008649308234453201, + "learning_rate": 0.0011806590754134636, + "loss": 3.9325060844421387, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.008581082336604595, + "learning_rate": 0.0011793605002092021, + "loss": 3.9200944900512695, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.008502699434757233, + "learning_rate": 0.0011780621954236238, + "loss": 3.9656100273132324, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.0083576999604702, + "learning_rate": 0.0011767641621027477, + "loss": 3.9130215644836426, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.008196269161999226, + "learning_rate": 0.0011754664012923785, + "loss": 3.942369222640991, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.008857307955622673, + "learning_rate": 0.0011741689140380956, + "loss": 3.909585952758789, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.008935349993407726, + "learning_rate": 0.0011728717013852635, + "loss": 3.9248366355895996, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.008821200579404831, + "learning_rate": 0.0011715747643790211, + "loss": 3.9589288234710693, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.00861942209303379, + "learning_rate": 0.0011702781040642885, + "loss": 3.9381728172302246, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.009127574972808361, + "learning_rate": 0.0011689817214857613, + "loss": 3.897829532623291, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.008008269593119621, + "learning_rate": 0.0011676856176879107, + "loss": 3.930056095123291, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.010070208460092545, + "learning_rate": 0.0011663897937149853, + "loss": 3.892270803451538, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.00903113093227148, + "learning_rate": 0.001165094250611006, + "loss": 3.9766440391540527, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.008872415870428085, + "learning_rate": 0.0011637989894197692, + "loss": 3.9559712409973145, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.009524816647171974, + "learning_rate": 0.0011625040111848427, + "loss": 3.9607882499694824, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.009836779907345772, + "learning_rate": 0.0011612093169495678, + "loss": 3.9629592895507812, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.009586540050804615, + "learning_rate": 0.0011599149077570561, + "loss": 3.95208740234375, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.010495901107788086, + "learning_rate": 0.0011586207846501896, + "loss": 3.9439730644226074, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.010346266441047192, + "learning_rate": 0.0011573269486716202, + "loss": 3.959299087524414, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.010767634026706219, + "learning_rate": 0.0011560334008637674, + "loss": 3.900784969329834, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.012122200801968575, + "learning_rate": 0.0011547401422688202, + "loss": 3.9588193893432617, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.012914645485579967, + "learning_rate": 0.0011534471739287331, + "loss": 3.9278407096862793, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.011785333044826984, + "learning_rate": 0.0011521544968852273, + "loss": 3.992793560028076, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.008455781266093254, + "learning_rate": 0.0011508621121797899, + "loss": 3.96525239944458, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.009655282832682133, + "learning_rate": 0.0011495700208536714, + "loss": 3.9346890449523926, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.010537004098296165, + "learning_rate": 0.0011482782239478871, + "loss": 3.9534668922424316, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.009438651613891125, + "learning_rate": 0.0011469867225032135, + "loss": 3.983487606048584, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.010586646385490894, + "learning_rate": 0.0011456955175601908, + "loss": 3.939680576324463, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.009082718752324581, + "learning_rate": 0.001144404610159119, + "loss": 3.9558515548706055, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.009833602234721184, + "learning_rate": 0.0011431140013400591, + "loss": 3.9539828300476074, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.009142170660197735, + "learning_rate": 0.0011418236921428313, + "loss": 3.9442849159240723, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.008603853173553944, + "learning_rate": 0.0011405336836070137, + "loss": 3.9526913166046143, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.008838302455842495, + "learning_rate": 0.0011392439767719446, + "loss": 3.973104953765869, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.008867006748914719, + "learning_rate": 0.0011379545726767156, + "loss": 3.9832305908203125, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.00971697736531496, + "learning_rate": 0.001136665472360178, + "loss": 3.9563937187194824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.00957175437361002, + "learning_rate": 0.001135376676860935, + "loss": 3.975194215774536, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.008548660203814507, + "learning_rate": 0.001134088187217347, + "loss": 3.9839985370635986, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.008556334301829338, + "learning_rate": 0.0011328000044675264, + "loss": 3.9225516319274902, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.008440867997705936, + "learning_rate": 0.0011315121296493385, + "loss": 3.945028066635132, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.009643333032727242, + "learning_rate": 0.0011302245638004017, + "loss": 3.943120002746582, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.008832835592329502, + "learning_rate": 0.001128937307958083, + "loss": 3.9895920753479004, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.008720474317669868, + "learning_rate": 0.0011276503631595036, + "loss": 3.941950798034668, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.008099926635622978, + "learning_rate": 0.001126363730441529, + "loss": 3.9626870155334473, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.008525107055902481, + "learning_rate": 0.0011250774108407778, + "loss": 3.950535774230957, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.008998272009193897, + "learning_rate": 0.0011237914053936135, + "loss": 3.9460678100585938, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.009387437254190445, + "learning_rate": 0.001122505715136148, + "loss": 3.926604986190796, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.007804017513990402, + "learning_rate": 0.0011212203411042386, + "loss": 3.9894230365753174, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.008564616553485394, + "learning_rate": 0.0011199352843334876, + "loss": 3.9339959621429443, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.008726440370082855, + "learning_rate": 0.0011186505458592427, + "loss": 3.9867987632751465, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.008074100129306316, + "learning_rate": 0.0011173661267165932, + "loss": 3.9492292404174805, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.007688041776418686, + "learning_rate": 0.0011160820279403744, + "loss": 3.9425454139709473, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.007808993570506573, + "learning_rate": 0.0011147982505651595, + "loss": 3.935114860534668, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.009786820039153099, + "learning_rate": 0.0011135147956252659, + "loss": 3.959758758544922, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.009487155824899673, + "learning_rate": 0.00111223166415475, + "loss": 3.989048480987549, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.008407804183661938, + "learning_rate": 0.0011109488571874073, + "loss": 3.932347297668457, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.007158380467444658, + "learning_rate": 0.0011096663757567723, + "loss": 3.9545154571533203, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.008840581402182579, + "learning_rate": 0.001108384220896117, + "loss": 3.965984344482422, + "step": 2160 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 9.208840083996672e+17, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-2160/training_args.bin b/runs/baseline/checkpoint-2160/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-2160/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-2520/chat_template.jinja b/runs/baseline/checkpoint-2520/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-2520/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-2520/config.json b/runs/baseline/checkpoint-2520/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-2520/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-2520/generation_config.json b/runs/baseline/checkpoint-2520/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-2520/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-2520/model.safetensors b/runs/baseline/checkpoint-2520/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d37cd51501d9c0f34cbe72c8a6bb76b4a4c16c43 --- /dev/null +++ b/runs/baseline/checkpoint-2520/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18ed9230704ff7b89d1d7eb34a36f2e139ce85cc503b3719af861f86780c0ae5 +size 583356232 diff --git a/runs/baseline/checkpoint-2520/optimizer.pt b/runs/baseline/checkpoint-2520/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..e131ed4259dedbf462877c2a1e1e1890168c22a0 --- /dev/null +++ b/runs/baseline/checkpoint-2520/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58a72143ae108aa9b217d96413039e371b321c8f03f7a20e6f3d746c03825e73 +size 1166825803 diff --git a/runs/baseline/checkpoint-2520/rng_state_0.pth b/runs/baseline/checkpoint-2520/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..f6a767b5a760715c982d5f3b7abf7f7ae7024858 --- /dev/null +++ b/runs/baseline/checkpoint-2520/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77a1aaa8594be4aeac424d02802142f8409ed1e2897c1e46e3c37ed12b93e83c +size 14917 diff --git a/runs/baseline/checkpoint-2520/rng_state_1.pth b/runs/baseline/checkpoint-2520/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..b0b2ce7324fa0e3d93e84f806452f72a932cb665 --- /dev/null +++ b/runs/baseline/checkpoint-2520/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8be9b03fd0264df00f58b952952d0e7adf408ae412c760c45b6919b81dd07768 +size 14917 diff --git a/runs/baseline/checkpoint-2520/scheduler.pt b/runs/baseline/checkpoint-2520/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..74bd0e9c9ee7711fa33882e1d3ae8d7f5d1dead2 --- /dev/null +++ b/runs/baseline/checkpoint-2520/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37e4964cd4dada742240c68b526bbf0c58037d9596fb0e4b5651dac33cc25e70 +size 1465 diff --git a/runs/baseline/checkpoint-2520/tokenizer.json b/runs/baseline/checkpoint-2520/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-2520/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-2520/tokenizer_config.json b/runs/baseline/checkpoint-2520/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-2520/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-2520/trainer_state.json b/runs/baseline/checkpoint-2520/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..15978b8ede1af2855045a833f74c9d75e90d9c80 --- /dev/null +++ b/runs/baseline/checkpoint-2520/trainer_state.json @@ -0,0 +1,17946 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 35.0, + "eval_steps": 500, + "global_step": 2520, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.641542434692383, + "eval_runtime": 51.6342, + "eval_samples_per_second": 47.294, + "eval_steps_per_second": 1.491, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.009694516658782959, + "learning_rate": 0.0020492498088767836, + "loss": 4.32808256149292, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.0097952950745821, + "learning_rate": 0.002048005909777114, + "loss": 4.280096054077148, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.010393359698355198, + "learning_rate": 0.0020467615812442106, + "loss": 4.317824363708496, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.00977372657507658, + "learning_rate": 0.0020455168242806054, + "loss": 4.3445329666137695, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.008568719029426575, + "learning_rate": 0.0020442716398891775, + "loss": 4.318026542663574, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.009732510894536972, + "learning_rate": 0.0020430260290731487, + "loss": 4.330609321594238, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.010429969057440758, + "learning_rate": 0.0020417799928360864, + "loss": 4.316794395446777, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.009292383678257465, + "learning_rate": 0.002040533532181898, + "loss": 4.317420959472656, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.011222455650568008, + "learning_rate": 0.0020392866481148356, + "loss": 4.292610168457031, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.013591725379228592, + "learning_rate": 0.0020380393416394894, + "loss": 4.325689315795898, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.013572026044130325, + "learning_rate": 0.002036791613760794, + "loss": 4.334960460662842, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.012986521236598492, + "learning_rate": 0.0020355434654840197, + "loss": 4.33829402923584, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.013321424834430218, + "learning_rate": 0.0020342948978147777, + "loss": 4.2926926612854, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.012307215481996536, + "learning_rate": 0.0020330459117590153, + "loss": 4.32757568359375, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.012909360229969025, + "learning_rate": 0.0020317965083230197, + "loss": 4.362889289855957, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017746610566973686, + "learning_rate": 0.0020305466885134115, + "loss": 4.3426313400268555, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.01696564070880413, + "learning_rate": 0.002029296453337148, + "loss": 4.3298516273498535, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.013348255306482315, + "learning_rate": 0.002028045803801522, + "loss": 4.360220432281494, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.013225145637989044, + "learning_rate": 0.002026794740914158, + "loss": 4.346250057220459, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.01569020375609398, + "learning_rate": 0.002025543265683016, + "loss": 4.308025360107422, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.018612321466207504, + "learning_rate": 0.002024291379116385, + "loss": 4.330190181732178, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.01619943603873253, + "learning_rate": 0.00202303908222289, + "loss": 4.339552402496338, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.015385213308036327, + "learning_rate": 0.0020217863760114826, + "loss": 4.3412184715271, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.018517112359404564, + "learning_rate": 0.002020533261491446, + "loss": 4.3289313316345215, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.014868509024381638, + "learning_rate": 0.0020192797396723903, + "loss": 4.359846115112305, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.014240248128771782, + "learning_rate": 0.0020180258115642578, + "loss": 4.369707107543945, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.018113158643245697, + "learning_rate": 0.0020167714781773143, + "loss": 4.362668514251709, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.01802460104227066, + "learning_rate": 0.002015516740522154, + "loss": 4.344742298126221, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.01414549071341753, + "learning_rate": 0.0020142615996096964, + "loss": 4.358080863952637, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.012157700955867767, + "learning_rate": 0.002013006056451186, + "loss": 4.347763538360596, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.011268534697592258, + "learning_rate": 0.0020117501120581905, + "loss": 4.335859298706055, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.010395628400146961, + "learning_rate": 0.002010493767442602, + "loss": 4.341609001159668, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.012147950939834118, + "learning_rate": 0.002009237023616635, + "loss": 4.328502655029297, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.00945813488215208, + "learning_rate": 0.0020079798815928252, + "loss": 4.311186790466309, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.009489500895142555, + "learning_rate": 0.0020067223423840294, + "loss": 4.371470928192139, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.009360029362142086, + "learning_rate": 0.0020054644070034234, + "loss": 4.33941125869751, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.010032779537141323, + "learning_rate": 0.0020042060764645037, + "loss": 4.365797519683838, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.009367046877741814, + "learning_rate": 0.002002947351781084, + "loss": 4.29443359375, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.008369733579456806, + "learning_rate": 0.0020016882339672967, + "loss": 4.294233322143555, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.010394785553216934, + "learning_rate": 0.0020004287240375893, + "loss": 4.352355003356934, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01068816613405943, + "learning_rate": 0.001999168823006728, + "loss": 4.342516899108887, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.009975973516702652, + "learning_rate": 0.0019979085318897905, + "loss": 4.346763610839844, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.009618151932954788, + "learning_rate": 0.001996647851702171, + "loss": 4.329115867614746, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.008399694226682186, + "learning_rate": 0.001995386783459577, + "loss": 4.28736686706543, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.008235268294811249, + "learning_rate": 0.001994125328178029, + "loss": 4.339872360229492, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.00836547277867794, + "learning_rate": 0.0019928634868738577, + "loss": 4.328156471252441, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.009619626216590405, + "learning_rate": 0.001991601260563706, + "loss": 4.337394714355469, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.009814038872718811, + "learning_rate": 0.0019903386502645275, + "loss": 4.325926780700684, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.009588954970240593, + "learning_rate": 0.001989075656993584, + "loss": 4.286477088928223, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.010471943765878677, + "learning_rate": 0.001987812281768447, + "loss": 4.322466850280762, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.010239035822451115, + "learning_rate": 0.001986548525606994, + "loss": 4.3487653732299805, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.008935568854212761, + "learning_rate": 0.001985284389527411, + "loss": 4.303601264953613, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.008624759502708912, + "learning_rate": 0.00198401987454819, + "loss": 4.34036922454834, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.009296547621488571, + "learning_rate": 0.0019827549816881277, + "loss": 4.273357391357422, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.009071138687431812, + "learning_rate": 0.0019814897119663248, + "loss": 4.321051597595215, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.008341389708220959, + "learning_rate": 0.0019802240664021876, + "loss": 4.343242645263672, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.007881715893745422, + "learning_rate": 0.0019789580460154227, + "loss": 4.32067346572876, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.008750553242862225, + "learning_rate": 0.0019776916518260404, + "loss": 4.330263614654541, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.008591460064053535, + "learning_rate": 0.001976424884854351, + "loss": 4.311957359313965, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.008084543980658054, + "learning_rate": 0.001975157746120966, + "loss": 4.331319808959961, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.008036594837903976, + "learning_rate": 0.001973890236646797, + "loss": 4.382656097412109, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.00805772002786398, + "learning_rate": 0.0019726223574530527, + "loss": 4.297247886657715, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.009019420482218266, + "learning_rate": 0.0019713541095612405, + "loss": 4.326526165008545, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.009953883476555347, + "learning_rate": 0.0019700854939931653, + "loss": 4.327356338500977, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.009340492077171803, + "learning_rate": 0.0019688165117709273, + "loss": 4.343993186950684, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.010395721532404423, + "learning_rate": 0.0019675471639169226, + "loss": 4.284934997558594, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.009946230798959732, + "learning_rate": 0.0019662774514538415, + "loss": 4.288936614990234, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.010981100611388683, + "learning_rate": 0.0019650073754046695, + "loss": 4.339935779571533, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.009688783437013626, + "learning_rate": 0.0019637369367926827, + "loss": 4.381425857543945, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.009442255832254887, + "learning_rate": 0.001962466136641452, + "loss": 4.304185390472412, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.009371291846036911, + "learning_rate": 0.0019611949759748364, + "loss": 4.306120872497559, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.009544814005494118, + "learning_rate": 0.001959923455816988, + "loss": 4.326251983642578, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.602745532989502, + "eval_runtime": 53.5041, + "eval_samples_per_second": 45.641, + "eval_steps_per_second": 1.439, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.01010364294052124, + "learning_rate": 0.0019586515771923474, + "loss": 4.24491024017334, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.012722729705274105, + "learning_rate": 0.0019573793411256446, + "loss": 4.257739543914795, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.01307204831391573, + "learning_rate": 0.0019561067486418967, + "loss": 4.291684150695801, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.011215966194868088, + "learning_rate": 0.0019548338007664098, + "loss": 4.237947463989258, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.012943019159138203, + "learning_rate": 0.0019535604985247727, + "loss": 4.242859363555908, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.01337282732129097, + "learning_rate": 0.0019522868429428636, + "loss": 4.238934516906738, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.013697991147637367, + "learning_rate": 0.0019510128350468438, + "loss": 4.285760402679443, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.012182159349322319, + "learning_rate": 0.001949738475863158, + "loss": 4.290044784545898, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.012620339170098305, + "learning_rate": 0.0019484637664185346, + "loss": 4.267187118530273, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.011284958571195602, + "learning_rate": 0.0019471887077399837, + "loss": 4.26662540435791, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.01161450706422329, + "learning_rate": 0.001945913300854797, + "loss": 4.283192157745361, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.010022968985140324, + "learning_rate": 0.001944637546790548, + "loss": 4.295843124389648, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.010341878980398178, + "learning_rate": 0.001943361446575087, + "loss": 4.295636177062988, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.011549185030162334, + "learning_rate": 0.0019420850012365457, + "loss": 4.288768768310547, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.012518810108304024, + "learning_rate": 0.0019408082118033331, + "loss": 4.229338645935059, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.012035328894853592, + "learning_rate": 0.0019395310793041353, + "loss": 4.258639335632324, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.011193782091140747, + "learning_rate": 0.0019382536047679146, + "loss": 4.239264488220215, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.009930136613547802, + "learning_rate": 0.0019369757892239096, + "loss": 4.252286434173584, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.00953229982405901, + "learning_rate": 0.0019356976337016327, + "loss": 4.293435096740723, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.010618340224027634, + "learning_rate": 0.0019344191392308716, + "loss": 4.235588550567627, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.008806651458144188, + "learning_rate": 0.0019331403068416856, + "loss": 4.26716423034668, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.008934796787798405, + "learning_rate": 0.0019318611375644065, + "loss": 4.258462905883789, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.009283283725380898, + "learning_rate": 0.0019305816324296386, + "loss": 4.257868766784668, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.009224812500178814, + "learning_rate": 0.0019293017924682555, + "loss": 4.258486747741699, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.008957969956099987, + "learning_rate": 0.0019280216187114016, + "loss": 4.309563636779785, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.010284051299095154, + "learning_rate": 0.0019267411121904894, + "loss": 4.275271415710449, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.009167762473225594, + "learning_rate": 0.0019254602739372001, + "loss": 4.238452911376953, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.009411636739969254, + "learning_rate": 0.0019241791049834816, + "loss": 4.2735595703125, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.009044171310961246, + "learning_rate": 0.0019228976063615497, + "loss": 4.27017879486084, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.008840068243443966, + "learning_rate": 0.0019216157791038834, + "loss": 4.28773307800293, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.009512359276413918, + "learning_rate": 0.001920333624243228, + "loss": 4.266008377075195, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.009946667589247227, + "learning_rate": 0.001919051142812593, + "loss": 4.279062271118164, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.00899894442409277, + "learning_rate": 0.0019177683358452505, + "loss": 4.277679443359375, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.00909668579697609, + "learning_rate": 0.0019164852043747338, + "loss": 4.311397552490234, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.009787327609956264, + "learning_rate": 0.0019152017494348408, + "loss": 4.30875825881958, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.008983899839222431, + "learning_rate": 0.001913917972059626, + "loss": 4.259649276733398, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.008411582559347153, + "learning_rate": 0.0019126338732834071, + "loss": 4.259811878204346, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.009162918664515018, + "learning_rate": 0.0019113494541407579, + "loss": 4.271515846252441, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.011304561980068684, + "learning_rate": 0.0019100647156665125, + "loss": 4.269217491149902, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.012483394704759121, + "learning_rate": 0.0019087796588957617, + "loss": 4.2949299812316895, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.012091380544006824, + "learning_rate": 0.0019074942848638522, + "loss": 4.289752006530762, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.010011567734181881, + "learning_rate": 0.0019062085946063869, + "loss": 4.294095993041992, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.008766217157244682, + "learning_rate": 0.0019049225891592226, + "loss": 4.256425857543945, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.012862488627433777, + "learning_rate": 0.001903636269558471, + "loss": 4.287847995758057, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.013917063362896442, + "learning_rate": 0.0019023496368404972, + "loss": 4.282784938812256, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.012129023671150208, + "learning_rate": 0.0019010626920419173, + "loss": 4.2486419677734375, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.010348426178097725, + "learning_rate": 0.0018997754361995986, + "loss": 4.289836883544922, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.011821693740785122, + "learning_rate": 0.001898487870350662, + "loss": 4.214344024658203, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.013136724010109901, + "learning_rate": 0.0018971999955324742, + "loss": 4.24208402633667, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.013173168525099754, + "learning_rate": 0.0018959118127826538, + "loss": 4.270164489746094, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.011018159799277782, + "learning_rate": 0.0018946233231390656, + "loss": 4.2571916580200195, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.01018658559769392, + "learning_rate": 0.0018933345276398227, + "loss": 4.283267498016357, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.009412688203155994, + "learning_rate": 0.0018920454273232848, + "loss": 4.272198677062988, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.007799267303198576, + "learning_rate": 0.0018907560232280557, + "loss": 4.260125637054443, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.008747464045882225, + "learning_rate": 0.0018894663163929863, + "loss": 4.2539215087890625, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.008003434166312218, + "learning_rate": 0.0018881763078571689, + "loss": 4.245024681091309, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.007593846879899502, + "learning_rate": 0.0018868859986599415, + "loss": 4.237929344177246, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.008099420927464962, + "learning_rate": 0.0018855953898408815, + "loss": 4.2789106369018555, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.008270486257970333, + "learning_rate": 0.0018843044824398097, + "loss": 4.263522148132324, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.009260379709303379, + "learning_rate": 0.0018830132774967866, + "loss": 4.275835037231445, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.009289492852985859, + "learning_rate": 0.0018817217760521132, + "loss": 4.313778400421143, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.009493921883404255, + "learning_rate": 0.0018804299791463281, + "loss": 4.303672790527344, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.009020023047924042, + "learning_rate": 0.0018791378878202105, + "loss": 4.245473384857178, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.009446556679904461, + "learning_rate": 0.0018778455031147724, + "loss": 4.267327785491943, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.009499974548816681, + "learning_rate": 0.0018765528260712672, + "loss": 4.301205635070801, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.008399393409490585, + "learning_rate": 0.00187525985773118, + "loss": 4.307802677154541, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.009519701823592186, + "learning_rate": 0.0018739665991362327, + "loss": 4.284796714782715, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01125232595950365, + "learning_rate": 0.0018726730513283804, + "loss": 4.285930633544922, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.011608537286520004, + "learning_rate": 0.00187137921534981, + "loss": 4.257689476013184, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.01064533181488514, + "learning_rate": 0.001870085092242944, + "loss": 4.249063968658447, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.009854599833488464, + "learning_rate": 0.0018687906830504321, + "loss": 4.255159378051758, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.01121739111840725, + "learning_rate": 0.0018674959888151576, + "loss": 4.287777900695801, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.5675368309021, + "eval_runtime": 51.8555, + "eval_samples_per_second": 47.092, + "eval_steps_per_second": 1.485, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.0104464590549469, + "learning_rate": 0.0018662010105802314, + "loss": 4.206080436706543, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.012800181284546852, + "learning_rate": 0.0018649057493889945, + "loss": 4.213550567626953, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.013382901437580585, + "learning_rate": 0.001863610206285015, + "loss": 4.190858840942383, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.014347050338983536, + "learning_rate": 0.0018623143823120894, + "loss": 4.22884464263916, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.01347152329981327, + "learning_rate": 0.001861018278514239, + "loss": 4.215579032897949, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.012786377221345901, + "learning_rate": 0.0018597218959357112, + "loss": 4.174999237060547, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.014137202873826027, + "learning_rate": 0.001858425235620979, + "loss": 4.21709680557251, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.013864399865269661, + "learning_rate": 0.001857128298614737, + "loss": 4.2375030517578125, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.010397658683359623, + "learning_rate": 0.0018558310859619048, + "loss": 4.231938362121582, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.00888819433748722, + "learning_rate": 0.0018545335987076223, + "loss": 4.195562839508057, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.010698067955672741, + "learning_rate": 0.001853235837897252, + "loss": 4.22847318649292, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.00987589918076992, + "learning_rate": 0.0018519378045763766, + "loss": 4.2147040367126465, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.011451782658696175, + "learning_rate": 0.001850639499790798, + "loss": 4.179263114929199, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.011224704794585705, + "learning_rate": 0.0018493409245865368, + "loss": 4.23030948638916, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.00996861606836319, + "learning_rate": 0.0018480420800098313, + "loss": 4.194283962249756, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.012143119238317013, + "learning_rate": 0.0018467429671071374, + "loss": 4.221704959869385, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.013126783072948456, + "learning_rate": 0.0018454435869251268, + "loss": 4.187748908996582, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.013031953014433384, + "learning_rate": 0.001844143940510687, + "loss": 4.248250961303711, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.012868993915617466, + "learning_rate": 0.001842844028910919, + "loss": 4.20161247253418, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.010102041065692902, + "learning_rate": 0.0018415438531731395, + "loss": 4.213654518127441, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.009599820710718632, + "learning_rate": 0.0018402434143448747, + "loss": 4.206313133239746, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.011536866426467896, + "learning_rate": 0.0018389427134738658, + "loss": 4.237072944641113, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.009558089077472687, + "learning_rate": 0.0018376417516080636, + "loss": 4.182195663452148, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01056219544261694, + "learning_rate": 0.0018363405297956302, + "loss": 4.232585906982422, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.012673584744334221, + "learning_rate": 0.0018350390490849354, + "loss": 4.228830337524414, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.011215884238481522, + "learning_rate": 0.0018337373105245598, + "loss": 4.216056823730469, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.011281915940344334, + "learning_rate": 0.00183243531516329, + "loss": 4.233404159545898, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.010381227359175682, + "learning_rate": 0.0018311330640501204, + "loss": 4.182013034820557, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.00900198519229889, + "learning_rate": 0.0018298305582342513, + "loss": 4.215488433837891, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.010198601521551609, + "learning_rate": 0.0018285277987650876, + "loss": 4.19389533996582, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.009965402074158192, + "learning_rate": 0.0018272247866922401, + "loss": 4.218005657196045, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.009299935773015022, + "learning_rate": 0.0018259215230655202, + "loss": 4.223684310913086, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.008977951481938362, + "learning_rate": 0.0018246180089349462, + "loss": 4.215778350830078, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.009261307306587696, + "learning_rate": 0.0018233142453507344, + "loss": 4.215066909790039, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.008322088047862053, + "learning_rate": 0.0018220102333633044, + "loss": 4.2188005447387695, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.008716144599020481, + "learning_rate": 0.0018207059740232739, + "loss": 4.187867641448975, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.008897020481526852, + "learning_rate": 0.0018194014683814625, + "loss": 4.218034267425537, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.007298174314200878, + "learning_rate": 0.0018180967174888857, + "loss": 4.2017011642456055, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.008491789922118187, + "learning_rate": 0.0018167917223967582, + "loss": 4.239873886108398, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.008413378149271011, + "learning_rate": 0.0018154864841564908, + "loss": 4.234771251678467, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.008141471073031425, + "learning_rate": 0.0018141810038196909, + "loss": 4.185944080352783, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.009480944834649563, + "learning_rate": 0.0018128752824381603, + "loss": 4.206586837768555, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.0106047997251153, + "learning_rate": 0.001811569321063894, + "loss": 4.235287666320801, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.010663514956831932, + "learning_rate": 0.0018102631207490838, + "loss": 4.204899311065674, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.011473523452877998, + "learning_rate": 0.001808956682546109, + "loss": 4.253668308258057, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01165992021560669, + "learning_rate": 0.0018076500075075456, + "loss": 4.209845542907715, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.011692410334944725, + "learning_rate": 0.0018063430966861564, + "loss": 4.250368595123291, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01061506848782301, + "learning_rate": 0.0018050359511348965, + "loss": 4.193821907043457, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.010510941036045551, + "learning_rate": 0.0018037285719069097, + "loss": 4.254327774047852, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.011587041430175304, + "learning_rate": 0.001802420960055528, + "loss": 4.2838945388793945, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.01257420051842928, + "learning_rate": 0.0018011131166342693, + "loss": 4.258241653442383, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.00936233438551426, + "learning_rate": 0.0017998050426968407, + "loss": 4.201854705810547, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.010093127377331257, + "learning_rate": 0.0017984967392971338, + "loss": 4.193051338195801, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.010032009333372116, + "learning_rate": 0.0017971882074892236, + "loss": 4.230212211608887, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.009455587714910507, + "learning_rate": 0.0017958794483273726, + "loss": 4.2157673835754395, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.009549238719046116, + "learning_rate": 0.001794570462866023, + "loss": 4.260272026062012, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.011059699580073357, + "learning_rate": 0.0017932612521598013, + "loss": 4.214818000793457, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.011653627268970013, + "learning_rate": 0.0017919518172635146, + "loss": 4.205112457275391, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.012438314035534859, + "learning_rate": 0.001790642159232151, + "loss": 4.179045677185059, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.011234673671424389, + "learning_rate": 0.0017893322791208784, + "loss": 4.239317417144775, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.010923845693469048, + "learning_rate": 0.0017880221779850434, + "loss": 4.230571269989014, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.009219309315085411, + "learning_rate": 0.001786711856880171, + "loss": 4.229068756103516, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.009849398396909237, + "learning_rate": 0.0017854013168619638, + "loss": 4.254772186279297, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.01117763202637434, + "learning_rate": 0.001784090558986299, + "loss": 4.252647399902344, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.010664612054824829, + "learning_rate": 0.0017827795843092316, + "loss": 4.204596519470215, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.00847950391471386, + "learning_rate": 0.0017814683938869895, + "loss": 4.211061954498291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.007305726408958435, + "learning_rate": 0.0017801569887759753, + "loss": 4.181221008300781, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.007970666512846947, + "learning_rate": 0.0017788453700327653, + "loss": 4.237855434417725, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.008873614482581615, + "learning_rate": 0.0017775335387141054, + "loss": 4.229188919067383, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.009691286832094193, + "learning_rate": 0.0017762214958769156, + "loss": 4.205491542816162, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.009360672906041145, + "learning_rate": 0.0017749092425782852, + "loss": 4.233466625213623, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.009895884431898594, + "learning_rate": 0.0017735967798754723, + "loss": 4.187763214111328, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.535672187805176, + "eval_runtime": 53.1736, + "eval_samples_per_second": 45.925, + "eval_steps_per_second": 1.448, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.00966129545122385, + "learning_rate": 0.0017722841088259045, + "loss": 4.1451735496521, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.010733380913734436, + "learning_rate": 0.0017709712304871774, + "loss": 4.1616363525390625, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.0117328567430377, + "learning_rate": 0.0017696581459170535, + "loss": 4.126331806182861, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.012640298344194889, + "learning_rate": 0.0017683448561734607, + "loss": 4.163730621337891, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.013074076734483242, + "learning_rate": 0.0017670313623144934, + "loss": 4.199502944946289, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.01132719125598669, + "learning_rate": 0.0017657176653984097, + "loss": 4.143380165100098, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.010008333250880241, + "learning_rate": 0.0017644037664836319, + "loss": 4.170320510864258, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.011090578511357307, + "learning_rate": 0.0017630896666287438, + "loss": 4.1571478843688965, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.011047456413507462, + "learning_rate": 0.0017617753668924924, + "loss": 4.101892948150635, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.010098653845489025, + "learning_rate": 0.0017604608683337857, + "loss": 4.154077529907227, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.010499860160052776, + "learning_rate": 0.00175914617201169, + "loss": 4.116868019104004, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.010939318686723709, + "learning_rate": 0.0017578312789854338, + "loss": 4.159144878387451, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.010622999630868435, + "learning_rate": 0.0017565161903144023, + "loss": 4.171411991119385, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.0108700692653656, + "learning_rate": 0.0017552009070581386, + "loss": 4.100748062133789, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.009794988669455051, + "learning_rate": 0.0017538854302763423, + "loss": 4.185161590576172, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.007750354241579771, + "learning_rate": 0.0017525697610288702, + "loss": 4.149083137512207, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.009923134930431843, + "learning_rate": 0.001751253900375732, + "loss": 4.205166816711426, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.010811141692101955, + "learning_rate": 0.0017499378493770936, + "loss": 4.131865978240967, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.010868688113987446, + "learning_rate": 0.0017486216090932733, + "loss": 4.148791790008545, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.011224126443266869, + "learning_rate": 0.001747305180584743, + "loss": 4.152503490447998, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.013262554071843624, + "learning_rate": 0.001745988564912124, + "loss": 4.164926528930664, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.013089599087834358, + "learning_rate": 0.001744671763136191, + "loss": 4.1626458168029785, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.01327646616846323, + "learning_rate": 0.0017433547763178665, + "loss": 4.189356803894043, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.014539042487740517, + "learning_rate": 0.0017420376055182233, + "loss": 4.176387786865234, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.013269098475575447, + "learning_rate": 0.001740720251798482, + "loss": 4.180716514587402, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.010922391898930073, + "learning_rate": 0.0017394027162200105, + "loss": 4.163928031921387, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.013355271890759468, + "learning_rate": 0.0017380849998443243, + "loss": 4.150012016296387, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.014149201102554798, + "learning_rate": 0.0017367671037330828, + "loss": 4.173713207244873, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01354879979044199, + "learning_rate": 0.0017354490289480918, + "loss": 4.150563716888428, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.013788704760372639, + "learning_rate": 0.0017341307765512998, + "loss": 4.213858604431152, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.011597222648561, + "learning_rate": 0.0017328123476047994, + "loss": 4.166989326477051, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.01055675558745861, + "learning_rate": 0.001731493743170824, + "loss": 4.172226428985596, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.011600397527217865, + "learning_rate": 0.0017301749643117506, + "loss": 4.1656951904296875, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.01114651933312416, + "learning_rate": 0.0017288560120900943, + "loss": 4.189825057983398, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.009811583906412125, + "learning_rate": 0.0017275368875685126, + "loss": 4.181164741516113, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.009758968837559223, + "learning_rate": 0.0017262175918097992, + "loss": 4.163188934326172, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.00999707542359829, + "learning_rate": 0.001724898125876887, + "loss": 4.184117317199707, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.009528180584311485, + "learning_rate": 0.001723578490832846, + "loss": 4.16407585144043, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.009717358276247978, + "learning_rate": 0.0017222586877408826, + "loss": 4.209892272949219, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.010909050703048706, + "learning_rate": 0.0017209387176643377, + "loss": 4.19977331161499, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.011234750039875507, + "learning_rate": 0.0017196185816666875, + "loss": 4.19363260269165, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.010300721041858196, + "learning_rate": 0.0017182982808115425, + "loss": 4.142427921295166, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.00941646471619606, + "learning_rate": 0.001716977816162644, + "loss": 4.159372329711914, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.009304054081439972, + "learning_rate": 0.0017156571887838679, + "loss": 4.207922458648682, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.009496180340647697, + "learning_rate": 0.0017143363997392184, + "loss": 4.137894630432129, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.009257436729967594, + "learning_rate": 0.0017130154500928326, + "loss": 4.198753833770752, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.00900586973875761, + "learning_rate": 0.0017116943409089746, + "loss": 4.160684585571289, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.009398494847118855, + "learning_rate": 0.00171037307325204, + "loss": 4.1819353103637695, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.008633013814687729, + "learning_rate": 0.0017090516481865483, + "loss": 4.176833152770996, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.00898425467312336, + "learning_rate": 0.0017077300667771484, + "loss": 4.190657138824463, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.008182398043572903, + "learning_rate": 0.0017064083300886155, + "loss": 4.173429489135742, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.009589143097400665, + "learning_rate": 0.0017050864391858478, + "loss": 4.1559014320373535, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.010304383002221584, + "learning_rate": 0.0017037643951338698, + "loss": 4.155453681945801, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.009943748824298382, + "learning_rate": 0.0017024421989978275, + "loss": 4.119734287261963, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.009238033555448055, + "learning_rate": 0.001701119851842992, + "loss": 4.188974857330322, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.008058690465986729, + "learning_rate": 0.0016997973547347531, + "loss": 4.19934606552124, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.008485953323543072, + "learning_rate": 0.0016984747087386233, + "loss": 4.165615558624268, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.008476605638861656, + "learning_rate": 0.0016971519149202347, + "loss": 4.21021842956543, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.008588708005845547, + "learning_rate": 0.0016958289743453385, + "loss": 4.208064556121826, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.009223178029060364, + "learning_rate": 0.0016945058880798037, + "loss": 4.158413887023926, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.009196167811751366, + "learning_rate": 0.0016931826571896166, + "loss": 4.203187465667725, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.009488921612501144, + "learning_rate": 0.001691859282740882, + "loss": 4.178227424621582, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.009253161028027534, + "learning_rate": 0.0016905357657998167, + "loss": 4.131230354309082, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.010069048963487148, + "learning_rate": 0.0016892121074327559, + "loss": 4.210734844207764, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.009945095516741276, + "learning_rate": 0.001687888308706147, + "loss": 4.184042930603027, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.010483640246093273, + "learning_rate": 0.0016865643706865505, + "loss": 4.156919479370117, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.011560715734958649, + "learning_rate": 0.0016852402944406391, + "loss": 4.185520172119141, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.010001277551054955, + "learning_rate": 0.0016839160810351978, + "loss": 4.226828575134277, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.010335934348404408, + "learning_rate": 0.0016825917315371202, + "loss": 4.183531284332275, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.011322776786983013, + "learning_rate": 0.001681267247013412, + "loss": 4.179635047912598, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.009663194417953491, + "learning_rate": 0.001679942628531186, + "loss": 4.165654182434082, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.010590246878564358, + "learning_rate": 0.0016786178771576643, + "loss": 4.1539459228515625, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.511386394500732, + "eval_runtime": 51.7751, + "eval_samples_per_second": 47.166, + "eval_steps_per_second": 1.487, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.01254354789853096, + "learning_rate": 0.0016772929939601728, + "loss": 4.104002475738525, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.01300987508147955, + "learning_rate": 0.0016759679800061483, + "loss": 4.127395153045654, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.013325816951692104, + "learning_rate": 0.0016746428363631299, + "loss": 4.131229400634766, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.014302870258688927, + "learning_rate": 0.001673317564098761, + "loss": 4.079358100891113, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.013776068575680256, + "learning_rate": 0.0016719921642807905, + "loss": 4.132065773010254, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.010788560844957829, + "learning_rate": 0.0016706666379770689, + "loss": 4.134626388549805, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.009766630828380585, + "learning_rate": 0.0016693409862555493, + "loss": 4.1645426750183105, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.011760029010474682, + "learning_rate": 0.0016680152101842842, + "loss": 4.110664367675781, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.011047836393117905, + "learning_rate": 0.0016666893108314288, + "loss": 4.129222869873047, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.009989139623939991, + "learning_rate": 0.0016653632892652344, + "loss": 4.076835632324219, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.010033783502876759, + "learning_rate": 0.001664037146554054, + "loss": 4.11555290222168, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.00996427796781063, + "learning_rate": 0.001662710883766337, + "loss": 4.131463527679443, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.01044023409485817, + "learning_rate": 0.0016613845019706288, + "loss": 4.098750114440918, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.010033872909843922, + "learning_rate": 0.0016600580022355712, + "loss": 4.161447048187256, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.009329567663371563, + "learning_rate": 0.0016587313856299012, + "loss": 4.146491527557373, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.010061589069664478, + "learning_rate": 0.0016574046532224504, + "loss": 4.107517719268799, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.009491910226643085, + "learning_rate": 0.0016560778060821424, + "loss": 4.098226547241211, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.008529345504939556, + "learning_rate": 0.0016547508452779943, + "loss": 4.1247711181640625, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.00936342403292656, + "learning_rate": 0.0016534237718791136, + "loss": 4.120573043823242, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.010255641303956509, + "learning_rate": 0.0016520965869547008, + "loss": 4.089067459106445, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.011638758704066277, + "learning_rate": 0.001650769291574044, + "loss": 4.074847221374512, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.014707857742905617, + "learning_rate": 0.0016494418868065214, + "loss": 4.153377532958984, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.01650111749768257, + "learning_rate": 0.0016481143737215986, + "loss": 4.134062767028809, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.014833372086286545, + "learning_rate": 0.00164678675338883, + "loss": 4.1344828605651855, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.012257886119186878, + "learning_rate": 0.0016454590268778538, + "loss": 4.098813533782959, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.013508322648704052, + "learning_rate": 0.0016441311952583959, + "loss": 4.106297492980957, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.01256900280714035, + "learning_rate": 0.001642803259600267, + "loss": 4.143631935119629, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.009885406121611595, + "learning_rate": 0.00164147522097336, + "loss": 4.120885848999023, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.012004674412310123, + "learning_rate": 0.0016401470804476518, + "loss": 4.089243412017822, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01224028691649437, + "learning_rate": 0.0016388188390932012, + "loss": 4.077765464782715, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.012511290609836578, + "learning_rate": 0.001637490497980149, + "loss": 4.161346435546875, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.010701716877520084, + "learning_rate": 0.0016361620581787143, + "loss": 4.10698938369751, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.010713113471865654, + "learning_rate": 0.001634833520759198, + "loss": 4.114283561706543, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.011018907651305199, + "learning_rate": 0.001633504886791978, + "loss": 4.134284019470215, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.011279584839940071, + "learning_rate": 0.0016321761573475117, + "loss": 4.1456804275512695, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.010642931796610355, + "learning_rate": 0.001630847333496332, + "loss": 4.160918235778809, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.011069678701460361, + "learning_rate": 0.0016295184163090482, + "loss": 4.136838912963867, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.011007166467607021, + "learning_rate": 0.0016281894068563443, + "loss": 4.129153728485107, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.010055703110992908, + "learning_rate": 0.0016268603062089802, + "loss": 4.138665676116943, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.009839230217039585, + "learning_rate": 0.0016255311154377875, + "loss": 4.1366496086120605, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.011191432364284992, + "learning_rate": 0.0016242018356136714, + "loss": 4.129880905151367, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.012130341492593288, + "learning_rate": 0.0016228724678076095, + "loss": 4.1425981521606445, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.012707795947790146, + "learning_rate": 0.0016215430130906482, + "loss": 4.115452766418457, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.012707809917628765, + "learning_rate": 0.001620213472533906, + "loss": 4.122758865356445, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.01182397548109293, + "learning_rate": 0.001618883847208569, + "loss": 4.140275955200195, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.010865340940654278, + "learning_rate": 0.001617554138185893, + "loss": 4.137759208679199, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.010973592288792133, + "learning_rate": 0.0016162243465371998, + "loss": 4.138947010040283, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.011247226968407631, + "learning_rate": 0.001614894473333879, + "loss": 4.115516662597656, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.010837487876415253, + "learning_rate": 0.0016135645196473853, + "loss": 4.160427093505859, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.009096396155655384, + "learning_rate": 0.0016122344865492385, + "loss": 4.1200032234191895, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.010671247728168964, + "learning_rate": 0.0016109043751110223, + "loss": 4.097919464111328, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.0107562355697155, + "learning_rate": 0.0016095741864043831, + "loss": 4.096290588378906, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.008909327909350395, + "learning_rate": 0.0016082439215010305, + "loss": 4.171017646789551, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.010305124334990978, + "learning_rate": 0.0016069135814727343, + "loss": 4.1450276374816895, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.011235613375902176, + "learning_rate": 0.0016055831673913255, + "loss": 4.134704113006592, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.00825632456690073, + "learning_rate": 0.0016042526803286949, + "loss": 4.107443332672119, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.009056378155946732, + "learning_rate": 0.0016029221213567921, + "loss": 4.132656097412109, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.008835164830088615, + "learning_rate": 0.001601591491547624, + "loss": 4.124729633331299, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.00915333442389965, + "learning_rate": 0.0016002607919732557, + "loss": 4.119500160217285, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.008532539941370487, + "learning_rate": 0.0015989300237058066, + "loss": 4.141661643981934, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.009952311404049397, + "learning_rate": 0.0015975991878174533, + "loss": 4.140929222106934, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.008920297026634216, + "learning_rate": 0.0015962682853804268, + "loss": 4.155877113342285, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.008770049549639225, + "learning_rate": 0.0015949373174670104, + "loss": 4.126394271850586, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.0078084105625748634, + "learning_rate": 0.0015936062851495411, + "loss": 4.128876209259033, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.00816641841083765, + "learning_rate": 0.0015922751895004078, + "loss": 4.137410640716553, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.008895459584891796, + "learning_rate": 0.0015909440315920505, + "loss": 4.07957649230957, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.008551528677344322, + "learning_rate": 0.001589612812496958, + "loss": 4.104213714599609, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.008325488306581974, + "learning_rate": 0.0015882815332876707, + "loss": 4.1584792137146, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.009341313503682613, + "learning_rate": 0.0015869501950367744, + "loss": 4.1417436599731445, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.00999050959944725, + "learning_rate": 0.0015856187988169067, + "loss": 4.136049270629883, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.008889932185411453, + "learning_rate": 0.0015842873457007484, + "loss": 4.16257381439209, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.009109573438763618, + "learning_rate": 0.0015829558367610268, + "loss": 4.171441555023193, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.485942840576172, + "eval_runtime": 52.1051, + "eval_samples_per_second": 46.867, + "eval_steps_per_second": 1.478, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.009881424717605114, + "learning_rate": 0.001581624273070515, + "loss": 4.059608459472656, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.011156721040606499, + "learning_rate": 0.0015802926557020294, + "loss": 4.047014236450195, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.012326992116868496, + "learning_rate": 0.0015789609857284309, + "loss": 4.06541109085083, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.01332208700478077, + "learning_rate": 0.0015776292642226213, + "loss": 4.042714595794678, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.012720128521323204, + "learning_rate": 0.0015762974922575446, + "loss": 4.098389625549316, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.010723215527832508, + "learning_rate": 0.0015749656709061854, + "loss": 4.059185981750488, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.010605281218886375, + "learning_rate": 0.001573633801241569, + "loss": 4.099538326263428, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.01101530622690916, + "learning_rate": 0.0015723018843367576, + "loss": 4.095674514770508, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.01135838869959116, + "learning_rate": 0.0015709699212648532, + "loss": 4.070535659790039, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.012119271792471409, + "learning_rate": 0.0015696379130989936, + "loss": 4.060605049133301, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.012783961370587349, + "learning_rate": 0.001568305860912355, + "loss": 4.075931072235107, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.011486220173537731, + "learning_rate": 0.0015669737657781461, + "loss": 4.075855255126953, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.01037552859634161, + "learning_rate": 0.0015656416287696136, + "loss": 4.087729454040527, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.009802743792533875, + "learning_rate": 0.0015643094509600354, + "loss": 4.0855607986450195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.010804438032209873, + "learning_rate": 0.001562977233422723, + "loss": 4.064528465270996, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.009505728259682655, + "learning_rate": 0.0015616449772310206, + "loss": 4.036775588989258, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.010634485632181168, + "learning_rate": 0.0015603126834583024, + "loss": 4.069923400878906, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.011392096057534218, + "learning_rate": 0.0015589803531779736, + "loss": 4.0812177658081055, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.011851722374558449, + "learning_rate": 0.0015576479874634683, + "loss": 4.09183406829834, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.011205444112420082, + "learning_rate": 0.0015563155873882502, + "loss": 4.07825231552124, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.009676349349319935, + "learning_rate": 0.0015549831540258093, + "loss": 4.052826404571533, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.009376165457069874, + "learning_rate": 0.0015536506884496638, + "loss": 4.02100133895874, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.00936955213546753, + "learning_rate": 0.0015523181917333564, + "loss": 4.049526214599609, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.00937784556299448, + "learning_rate": 0.0015509856649504564, + "loss": 4.096356391906738, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.010014008730649948, + "learning_rate": 0.0015496531091745555, + "loss": 4.03254508972168, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.009909634478390217, + "learning_rate": 0.0015483205254792705, + "loss": 4.067630767822266, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.009319501928985119, + "learning_rate": 0.0015469879149382399, + "loss": 4.094167709350586, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.008634721860289574, + "learning_rate": 0.0015456552786251244, + "loss": 4.122404098510742, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.009412454441189766, + "learning_rate": 0.0015443226176136049, + "loss": 4.081546783447266, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.010102108120918274, + "learning_rate": 0.0015429899329773814, + "loss": 4.049839019775391, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.010921654291450977, + "learning_rate": 0.0015416572257901746, + "loss": 4.097097396850586, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.010607713833451271, + "learning_rate": 0.0015403244971257226, + "loss": 4.084481239318848, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.011887960135936737, + "learning_rate": 0.0015389917480577806, + "loss": 4.074968338012695, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01089888159185648, + "learning_rate": 0.0015376589796601203, + "loss": 4.062132835388184, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.009934183210134506, + "learning_rate": 0.0015363261930065298, + "loss": 4.128092288970947, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.00878534372895956, + "learning_rate": 0.0015349933891708105, + "loss": 4.069051742553711, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.009928155690431595, + "learning_rate": 0.0015336605692267788, + "loss": 4.073912620544434, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.011091411113739014, + "learning_rate": 0.0015323277342482637, + "loss": 4.125944137573242, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.010314172133803368, + "learning_rate": 0.0015309948853091055, + "loss": 4.062620162963867, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.010423967614769936, + "learning_rate": 0.001529662023483158, + "loss": 4.101737976074219, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.009822997264564037, + "learning_rate": 0.0015283291498442817, + "loss": 4.064520359039307, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.009757095016539097, + "learning_rate": 0.001526996265466351, + "loss": 4.100624084472656, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.008650688454508781, + "learning_rate": 0.001525663371423246, + "loss": 4.109685897827148, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.007378234062343836, + "learning_rate": 0.0015243304687888553, + "loss": 4.044726848602295, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.008194336667656898, + "learning_rate": 0.001522997558637074, + "loss": 4.097923278808594, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.009764892049133778, + "learning_rate": 0.0015216646420418048, + "loss": 4.102148056030273, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.009242868050932884, + "learning_rate": 0.001520331720076953, + "loss": 4.068563461303711, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.010409324429929256, + "learning_rate": 0.001518998793816431, + "loss": 4.065944194793701, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.009215258061885834, + "learning_rate": 0.0015176658643341527, + "loss": 4.10223388671875, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.008065393194556236, + "learning_rate": 0.0015163329327040366, + "loss": 4.099432945251465, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.008498288691043854, + "learning_rate": 0.001515, + "loss": 4.109854698181152, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.008919150568544865, + "learning_rate": 0.0015136670672959633, + "loss": 4.128247261047363, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.009104429744184017, + "learning_rate": 0.0015123341356658474, + "loss": 4.068622589111328, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.008558365516364574, + "learning_rate": 0.0015110012061835692, + "loss": 4.114214897155762, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.007865076884627342, + "learning_rate": 0.0015096682799230477, + "loss": 4.11155366897583, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.007848534733057022, + "learning_rate": 0.0015083353579581958, + "loss": 4.045694351196289, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.008238030597567558, + "learning_rate": 0.0015070024413629264, + "loss": 4.093930721282959, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.007536398712545633, + "learning_rate": 0.0015056695312111448, + "loss": 4.106235504150391, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.008156892843544483, + "learning_rate": 0.001504336628576754, + "loss": 4.13060998916626, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.008144439198076725, + "learning_rate": 0.0015030037345336487, + "loss": 4.091304779052734, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.007757358253002167, + "learning_rate": 0.001501670850155718, + "loss": 4.069276809692383, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.00860623549669981, + "learning_rate": 0.0015003379765168425, + "loss": 4.107658386230469, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.008614457212388515, + "learning_rate": 0.0014990051146908949, + "loss": 4.106705665588379, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.009120592847466469, + "learning_rate": 0.001497672265751737, + "loss": 4.0891923904418945, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.009446980431675911, + "learning_rate": 0.0014963394307732213, + "loss": 4.098847389221191, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.010494766756892204, + "learning_rate": 0.0014950066108291894, + "loss": 4.115325450897217, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.010057658888399601, + "learning_rate": 0.0014936738069934706, + "loss": 4.099984169006348, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.009071771055459976, + "learning_rate": 0.0014923410203398796, + "loss": 4.091543674468994, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.009937075898051262, + "learning_rate": 0.00149100825194222, + "loss": 4.086492538452148, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.010222986340522766, + "learning_rate": 0.0014896755028742778, + "loss": 4.063841819763184, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.010513762012124062, + "learning_rate": 0.001488342774209826, + "loss": 4.11991548538208, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.009530222974717617, + "learning_rate": 0.0014870100670226192, + "loss": 4.043334007263184, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.467459678649902, + "eval_runtime": 54.1628, + "eval_samples_per_second": 45.086, + "eval_steps_per_second": 1.422, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.010274187661707401, + "learning_rate": 0.0014856773823863955, + "loss": 4.049561500549316, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.011448442004621029, + "learning_rate": 0.001484344721374876, + "loss": 4.019484996795654, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.011361192911863327, + "learning_rate": 0.0014830120850617599, + "loss": 4.0429887771606445, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.010250156745314598, + "learning_rate": 0.0014816794745207299, + "loss": 4.012682914733887, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.010024400427937508, + "learning_rate": 0.0014803468908254449, + "loss": 3.979409694671631, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.01153535582125187, + "learning_rate": 0.0014790143350495446, + "loss": 4.045780181884766, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.012632817961275578, + "learning_rate": 0.001477681808266644, + "loss": 4.057641983032227, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.013758293353021145, + "learning_rate": 0.0014763493115503366, + "loss": 4.061607360839844, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.014041583053767681, + "learning_rate": 0.001475016845974191, + "loss": 4.035029411315918, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.013629456982016563, + "learning_rate": 0.0014736844126117497, + "loss": 4.062169075012207, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.011023213155567646, + "learning_rate": 0.0014723520125365323, + "loss": 4.02731990814209, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.01001750584691763, + "learning_rate": 0.0014710196468220266, + "loss": 4.019882678985596, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.008321646600961685, + "learning_rate": 0.001469687316541698, + "loss": 4.005446434020996, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.008935946971178055, + "learning_rate": 0.0014683550227689796, + "loss": 4.017102241516113, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.009137130342423916, + "learning_rate": 0.001467022766577277, + "loss": 4.024102210998535, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.00805022194981575, + "learning_rate": 0.001465690549039965, + "loss": 3.9935250282287598, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.009204058907926083, + "learning_rate": 0.0014643583712303865, + "loss": 4.042818546295166, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.009812815114855766, + "learning_rate": 0.001463026234221854, + "loss": 4.0252485275268555, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.009955782443284988, + "learning_rate": 0.0014616941390876454, + "loss": 3.9839768409729004, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.009374645538628101, + "learning_rate": 0.001460362086901007, + "loss": 4.040245056152344, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.010008739307522774, + "learning_rate": 0.0014590300787351471, + "loss": 4.0306596755981445, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.009014200419187546, + "learning_rate": 0.001457698115663243, + "loss": 4.028713226318359, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.008381523191928864, + "learning_rate": 0.0014563661987584313, + "loss": 4.024718284606934, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.008627948351204395, + "learning_rate": 0.0014550343290938143, + "loss": 4.02303409576416, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.010257705114781857, + "learning_rate": 0.0014537025077424556, + "loss": 4.061058044433594, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.010759209282696247, + "learning_rate": 0.0014523707357773788, + "loss": 4.064718246459961, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.009966139681637287, + "learning_rate": 0.0014510390142715697, + "loss": 3.9919447898864746, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.00997584406286478, + "learning_rate": 0.001449707344297971, + "loss": 4.0431599617004395, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.010846944525837898, + "learning_rate": 0.0014483757269294858, + "loss": 4.0222697257995605, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.00991352554410696, + "learning_rate": 0.0014470441632389735, + "loss": 4.059438705444336, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.008994397707283497, + "learning_rate": 0.0014457126542992518, + "loss": 4.021494388580322, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.009567267261445522, + "learning_rate": 0.0014443812011830934, + "loss": 4.070810317993164, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.009199011139571667, + "learning_rate": 0.0014430498049632255, + "loss": 4.020938873291016, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.010440350510179996, + "learning_rate": 0.0014417184667123303, + "loss": 4.062529563903809, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.009962482377886772, + "learning_rate": 0.0014403871875030424, + "loss": 4.0687150955200195, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.009389379993081093, + "learning_rate": 0.0014390559684079502, + "loss": 4.055185794830322, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.011469941586256027, + "learning_rate": 0.0014377248104995924, + "loss": 4.041023254394531, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.011034650728106499, + "learning_rate": 0.0014363937148504588, + "loss": 4.054254531860352, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.010651151649653912, + "learning_rate": 0.0014350626825329897, + "loss": 4.037753582000732, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.010747930034995079, + "learning_rate": 0.0014337317146195733, + "loss": 4.059938907623291, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.00987576600164175, + "learning_rate": 0.0014324008121825468, + "loss": 4.075366973876953, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.008504371158778667, + "learning_rate": 0.0014310699762941936, + "loss": 4.026755332946777, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.009445052593946457, + "learning_rate": 0.001429739208026745, + "loss": 4.034431457519531, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.010560023598372936, + "learning_rate": 0.0014284085084523763, + "loss": 4.101158618927002, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.009920580312609673, + "learning_rate": 0.0014270778786432078, + "loss": 4.041321754455566, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.009854268282651901, + "learning_rate": 0.0014257473196713053, + "loss": 4.049882411956787, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.009825135581195354, + "learning_rate": 0.0014244168326086745, + "loss": 4.067440986633301, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.010506784543395042, + "learning_rate": 0.0014230864185272663, + "loss": 4.042476654052734, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.009527767077088356, + "learning_rate": 0.0014217560784989696, + "loss": 4.055248260498047, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.01013826485723257, + "learning_rate": 0.001420425813595617, + "loss": 4.05246639251709, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.011020111851394176, + "learning_rate": 0.0014190956248889778, + "loss": 4.032987117767334, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.010647117160260677, + "learning_rate": 0.0014177655134507614, + "loss": 4.093589782714844, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.009484780952334404, + "learning_rate": 0.001416435480352615, + "loss": 4.017437934875488, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.0114199910312891, + "learning_rate": 0.0014151055266661211, + "loss": 4.080643653869629, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.011137664318084717, + "learning_rate": 0.0014137756534628007, + "loss": 4.069380760192871, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.008972682990133762, + "learning_rate": 0.0014124458618141073, + "loss": 4.051383018493652, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.010311643593013287, + "learning_rate": 0.0014111161527914317, + "loss": 4.048953533172607, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.01113453321158886, + "learning_rate": 0.0014097865274660944, + "loss": 4.063544273376465, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.01091703586280346, + "learning_rate": 0.0014084569869093518, + "loss": 4.006721496582031, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.010672449134290218, + "learning_rate": 0.0014071275321923906, + "loss": 4.020220756530762, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.009682527743279934, + "learning_rate": 0.001405798164386328, + "loss": 4.057307243347168, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.009808676317334175, + "learning_rate": 0.0014044688845622126, + "loss": 4.105412483215332, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.010351189412176609, + "learning_rate": 0.0014031396937910202, + "loss": 4.066454887390137, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.010038275271654129, + "learning_rate": 0.001401810593143656, + "loss": 4.056752681732178, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01016094908118248, + "learning_rate": 0.001400481583690952, + "loss": 4.097933292388916, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.008683547377586365, + "learning_rate": 0.0013991526665036679, + "loss": 4.0386152267456055, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.009556332603096962, + "learning_rate": 0.0013978238426524882, + "loss": 4.05835485458374, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.009597044438123703, + "learning_rate": 0.001396495113208022, + "loss": 4.052363872528076, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.010021139867603779, + "learning_rate": 0.0013951664792408026, + "loss": 4.083431243896484, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.010611358098685741, + "learning_rate": 0.0013938379418212863, + "loss": 4.042330265045166, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.009132112376391888, + "learning_rate": 0.001392509502019852, + "loss": 4.058825969696045, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.009808709844946861, + "learning_rate": 0.0013911811609067992, + "loss": 4.093010902404785, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.4473161697387695, + "eval_runtime": 51.7738, + "eval_samples_per_second": 47.167, + "eval_steps_per_second": 1.487, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.011859052814543247, + "learning_rate": 0.0013898529195523488, + "loss": 3.9572651386260986, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.013210238888859749, + "learning_rate": 0.0013885247790266406, + "loss": 3.9812350273132324, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.012981324456632137, + "learning_rate": 0.001387196740399733, + "loss": 3.987725019454956, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.01424382347613573, + "learning_rate": 0.001385868804741604, + "loss": 3.9719889163970947, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.0154642965644598, + "learning_rate": 0.0013845409731221465, + "loss": 3.984248161315918, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.014779271557927132, + "learning_rate": 0.0013832132466111706, + "loss": 4.002912521362305, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.012921188026666641, + "learning_rate": 0.0013818856262784013, + "loss": 3.9588499069213867, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01165325753390789, + "learning_rate": 0.0013805581131934792, + "loss": 3.993288993835449, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.008908954448997974, + "learning_rate": 0.0013792307084259563, + "loss": 3.987151861190796, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.012040507979691029, + "learning_rate": 0.0013779034130452991, + "loss": 3.9675703048706055, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.013464349322021008, + "learning_rate": 0.0013765762281208868, + "loss": 3.9901437759399414, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.009626250714063644, + "learning_rate": 0.0013752491547220063, + "loss": 3.9373281002044678, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.00897257961332798, + "learning_rate": 0.0013739221939178582, + "loss": 4.00804328918457, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.010689414106309414, + "learning_rate": 0.00137259534677755, + "loss": 4.020804405212402, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.010462156496942043, + "learning_rate": 0.0013712686143700988, + "loss": 4.028973579406738, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.009135857224464417, + "learning_rate": 0.001369941997764429, + "loss": 3.9798216819763184, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.009298899210989475, + "learning_rate": 0.0013686154980293713, + "loss": 4.018845081329346, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.00979928020387888, + "learning_rate": 0.0013672891162336636, + "loss": 3.9946775436401367, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.008175542578101158, + "learning_rate": 0.0013659628534459457, + "loss": 3.981858491897583, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.008302642032504082, + "learning_rate": 0.0013646367107347662, + "loss": 3.991128444671631, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.008516009896993637, + "learning_rate": 0.0013633106891685718, + "loss": 4.004990100860596, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.007936482317745686, + "learning_rate": 0.0013619847898157164, + "loss": 3.995734214782715, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.008317600004374981, + "learning_rate": 0.001360659013744451, + "loss": 4.040653228759766, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.008481282740831375, + "learning_rate": 0.001359333362022931, + "loss": 3.969860553741455, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.008944948203861713, + "learning_rate": 0.0013580078357192096, + "loss": 3.9908909797668457, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.008897850289940834, + "learning_rate": 0.001356682435901239, + "loss": 4.04786491394043, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.009527456946671009, + "learning_rate": 0.001355357163636871, + "loss": 4.017423629760742, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.011714582331478596, + "learning_rate": 0.001354032019993852, + "loss": 3.9990692138671875, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.011083807796239853, + "learning_rate": 0.0013527070060398274, + "loss": 3.995265483856201, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.009878288954496384, + "learning_rate": 0.0013513821228423363, + "loss": 4.0364532470703125, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.009727329947054386, + "learning_rate": 0.0013500573714688136, + "loss": 4.035953521728516, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.009482033550739288, + "learning_rate": 0.0013487327529865879, + "loss": 4.015855312347412, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.009037062525749207, + "learning_rate": 0.0013474082684628798, + "loss": 4.024164199829102, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.009094092063605785, + "learning_rate": 0.0013460839189648028, + "loss": 4.037765979766846, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.008230654522776604, + "learning_rate": 0.0013447597055593612, + "loss": 4.006317615509033, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.008491289801895618, + "learning_rate": 0.0013434356293134501, + "loss": 4.042603969573975, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.008907693438231945, + "learning_rate": 0.0013421116912938533, + "loss": 4.008098602294922, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.008897818624973297, + "learning_rate": 0.0013407878925672438, + "loss": 3.9892945289611816, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.00957555789500475, + "learning_rate": 0.0013394642342001832, + "loss": 4.046009063720703, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.009948568418622017, + "learning_rate": 0.0013381407172591185, + "loss": 3.9760642051696777, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.01014955248683691, + "learning_rate": 0.0013368173428103835, + "loss": 4.023215293884277, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.008578895591199398, + "learning_rate": 0.0013354941119201966, + "loss": 4.002819061279297, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.008729875087738037, + "learning_rate": 0.001334171025654662, + "loss": 4.0340657234191895, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.009556280449032784, + "learning_rate": 0.0013328480850797655, + "loss": 3.983076333999634, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.009733820334076881, + "learning_rate": 0.0013315252912613766, + "loss": 4.066578388214111, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.009025252424180508, + "learning_rate": 0.0013302026452652474, + "loss": 3.9980697631835938, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.008485320024192333, + "learning_rate": 0.001328880148157008, + "loss": 4.018826961517334, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.009539696387946606, + "learning_rate": 0.0013275578010021726, + "loss": 4.011737823486328, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.010881016030907631, + "learning_rate": 0.0013262356048661305, + "loss": 4.007291793823242, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.010362236760556698, + "learning_rate": 0.0013249135608141523, + "loss": 3.9929394721984863, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.010296608321368694, + "learning_rate": 0.0013235916699113849, + "loss": 4.0015130043029785, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.011211174540221691, + "learning_rate": 0.0013222699332228517, + "loss": 4.005324363708496, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.009743994101881981, + "learning_rate": 0.0013209483518134525, + "loss": 4.0278143882751465, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.008563914336264133, + "learning_rate": 0.0013196269267479607, + "loss": 3.9997801780700684, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.009078902192413807, + "learning_rate": 0.0013183056590910255, + "loss": 4.026200294494629, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.009618895128369331, + "learning_rate": 0.0013169845499071675, + "loss": 4.027050018310547, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.008920644409954548, + "learning_rate": 0.0013156636002607822, + "loss": 4.000490665435791, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.008849202655255795, + "learning_rate": 0.0013143428112161323, + "loss": 4.044985294342041, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.009303621016442776, + "learning_rate": 0.001313022183837356, + "loss": 4.046281814575195, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.01076227892190218, + "learning_rate": 0.0013117017191884579, + "loss": 4.030576705932617, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.010400841943919659, + "learning_rate": 0.0013103814183333127, + "loss": 3.980081558227539, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.009459166787564754, + "learning_rate": 0.001309061282335663, + "loss": 4.014824867248535, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.009054969064891338, + "learning_rate": 0.001307741312259118, + "loss": 4.0485944747924805, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.009003991261124611, + "learning_rate": 0.0013064215091671545, + "loss": 3.9990320205688477, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.008959448896348476, + "learning_rate": 0.001305101874123113, + "loss": 4.038620948791504, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.009102243930101395, + "learning_rate": 0.0013037824081902015, + "loss": 4.030906677246094, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.008599337190389633, + "learning_rate": 0.0013024631124314877, + "loss": 4.02648401260376, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.008771977387368679, + "learning_rate": 0.0013011439879099056, + "loss": 4.027151107788086, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.009082595817744732, + "learning_rate": 0.00129982503568825, + "loss": 4.0194478034973145, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.009025354869663715, + "learning_rate": 0.0012985062568291763, + "loss": 4.044382095336914, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.00875102635473013, + "learning_rate": 0.0012971876523952014, + "loss": 4.0257568359375, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.009312704205513, + "learning_rate": 0.0012958692234487006, + "loss": 3.9964494705200195, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.431707859039307, + "eval_runtime": 50.7311, + "eval_samples_per_second": 48.136, + "eval_steps_per_second": 1.518, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.00992550328373909, + "learning_rate": 0.0012945509710519088, + "loss": 3.9253792762756348, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.009920453652739525, + "learning_rate": 0.0012932328962669173, + "loss": 3.9438819885253906, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.009437218308448792, + "learning_rate": 0.0012919150001556756, + "loss": 3.960963249206543, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.009616893716156483, + "learning_rate": 0.0012905972837799896, + "loss": 3.969301223754883, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.009107022546231747, + "learning_rate": 0.0012892797482015185, + "loss": 3.9626078605651855, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.011317167431116104, + "learning_rate": 0.0012879623944817772, + "loss": 3.946483850479126, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.010536770336329937, + "learning_rate": 0.001286645223682134, + "loss": 3.959566116333008, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.009349033236503601, + "learning_rate": 0.0012853282368638097, + "loss": 3.964435338973999, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.009950080886483192, + "learning_rate": 0.0012840114350878762, + "loss": 3.9729413986206055, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.010833023115992546, + "learning_rate": 0.001282694819415257, + "loss": 4.006421089172363, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.011538525111973286, + "learning_rate": 0.0012813783909067266, + "loss": 3.9558355808258057, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.010671315714716911, + "learning_rate": 0.0012800621506229065, + "loss": 4.0052337646484375, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.010835555382072926, + "learning_rate": 0.0012787460996242686, + "loss": 3.958641767501831, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.01082367543131113, + "learning_rate": 0.0012774302389711304, + "loss": 3.9818902015686035, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.009194346144795418, + "learning_rate": 0.0012761145697236583, + "loss": 3.9688854217529297, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.007859935984015465, + "learning_rate": 0.001274799092941862, + "loss": 3.9730639457702637, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.008534743450582027, + "learning_rate": 0.0012734838096855978, + "loss": 3.975003719329834, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.009369445964694023, + "learning_rate": 0.0012721687210145666, + "loss": 3.972907304763794, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.0090147340670228, + "learning_rate": 0.0012708538279883098, + "loss": 3.964256525039673, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.007745139300823212, + "learning_rate": 0.0012695391316662151, + "loss": 3.9504833221435547, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.008920101448893547, + "learning_rate": 0.0012682246331075075, + "loss": 3.9735097885131836, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.008987226523458958, + "learning_rate": 0.0012669103333712568, + "loss": 3.9482927322387695, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.00841289572417736, + "learning_rate": 0.0012655962335163685, + "loss": 3.965221881866455, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.008016116917133331, + "learning_rate": 0.0012642823346015902, + "loss": 3.9390206336975098, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.008470152504742146, + "learning_rate": 0.0012629686376855068, + "loss": 3.9257445335388184, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.008637234568595886, + "learning_rate": 0.0012616551438265394, + "loss": 3.9999356269836426, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.008036923594772816, + "learning_rate": 0.0012603418540829473, + "loss": 3.9755172729492188, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.008908843621611595, + "learning_rate": 0.0012590287695128232, + "loss": 3.9248545169830322, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.008743193000555038, + "learning_rate": 0.001257715891174096, + "loss": 3.9619390964508057, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.00834199320524931, + "learning_rate": 0.001256403220124528, + "loss": 3.961569309234619, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.009148893877863884, + "learning_rate": 0.001255090757421715, + "loss": 3.972104549407959, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.00925084762275219, + "learning_rate": 0.0012537785041230843, + "loss": 4.000775337219238, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.008659154176712036, + "learning_rate": 0.0012524664612858947, + "loss": 3.954564332962036, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.007787859998643398, + "learning_rate": 0.0012511546299672355, + "loss": 3.9498860836029053, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.00817918125540018, + "learning_rate": 0.0012498430112240246, + "loss": 4.003897666931152, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.009282548911869526, + "learning_rate": 0.001248531606113011, + "loss": 3.9621925354003906, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.009375610388815403, + "learning_rate": 0.001247220415690769, + "loss": 3.9977524280548096, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.009650378488004208, + "learning_rate": 0.001245909441013701, + "loss": 3.9882829189300537, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.009350919164717197, + "learning_rate": 0.0012445986831380366, + "loss": 4.019596099853516, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.008726303465664387, + "learning_rate": 0.0012432881431198289, + "loss": 3.9933621883392334, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.00829484686255455, + "learning_rate": 0.0012419778220149568, + "loss": 4.001237869262695, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.008744452148675919, + "learning_rate": 0.0012406677208791218, + "loss": 3.9323887825012207, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.008170492947101593, + "learning_rate": 0.0012393578407678496, + "loss": 3.925581455230713, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.007843422703444958, + "learning_rate": 0.001238048182736486, + "loss": 4.001023769378662, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.008914309553802013, + "learning_rate": 0.0012367387478401987, + "loss": 3.9804537296295166, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.009736751206219196, + "learning_rate": 0.0012354295371339774, + "loss": 3.9831767082214355, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.010061381384730339, + "learning_rate": 0.0012341205516726273, + "loss": 3.9462435245513916, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.010350692085921764, + "learning_rate": 0.0012328117925107765, + "loss": 4.012863636016846, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01039902213960886, + "learning_rate": 0.0012315032607028667, + "loss": 3.9811301231384277, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.011423371732234955, + "learning_rate": 0.0012301949573031595, + "loss": 3.9773542881011963, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.012449614703655243, + "learning_rate": 0.001228886883365731, + "loss": 3.969600200653076, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.010361429303884506, + "learning_rate": 0.0012275790399444726, + "loss": 4.035360336303711, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.009603082202374935, + "learning_rate": 0.0012262714280930904, + "loss": 3.9840846061706543, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.009183845482766628, + "learning_rate": 0.0012249640488651036, + "loss": 4.031542778015137, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.009694446809589863, + "learning_rate": 0.0012236569033138441, + "loss": 3.9935057163238525, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.0097816726192832, + "learning_rate": 0.0012223499924924545, + "loss": 3.9851176738739014, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.009413184598088264, + "learning_rate": 0.001221043317453891, + "loss": 3.9642205238342285, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.009960711002349854, + "learning_rate": 0.0012197368792509166, + "loss": 3.9373817443847656, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.010124670341610909, + "learning_rate": 0.0012184306789361059, + "loss": 3.958115577697754, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.011087656021118164, + "learning_rate": 0.00121712471756184, + "loss": 3.9607057571411133, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.009812800213694572, + "learning_rate": 0.001215818996180309, + "loss": 4.00722599029541, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.010305573232471943, + "learning_rate": 0.0012145135158435094, + "loss": 3.98728084564209, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.01019787136465311, + "learning_rate": 0.0012132082776032422, + "loss": 3.981172800064087, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.009260181337594986, + "learning_rate": 0.001211903282511115, + "loss": 4.034339904785156, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.00972998607903719, + "learning_rate": 0.001210598531618538, + "loss": 4.022885322570801, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.009853586554527283, + "learning_rate": 0.0012092940259767267, + "loss": 4.000235080718994, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.00961449183523655, + "learning_rate": 0.0012079897666366962, + "loss": 3.9951353073120117, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.010257176123559475, + "learning_rate": 0.0012066857546492657, + "loss": 3.9801371097564697, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.009443561546504498, + "learning_rate": 0.001205381991065054, + "loss": 3.9596502780914307, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.010389018803834915, + "learning_rate": 0.0012040784769344795, + "loss": 3.980527639389038, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.010322567075490952, + "learning_rate": 0.0012027752133077607, + "loss": 3.996258497238159, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.009987728670239449, + "learning_rate": 0.0012014722012349125, + "loss": 3.9493203163146973, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.421086311340332, + "eval_runtime": 52.5623, + "eval_samples_per_second": 46.459, + "eval_steps_per_second": 1.465, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.009405115619301796, + "learning_rate": 0.0012001694417657493, + "loss": 3.927506923675537, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.009974643588066101, + "learning_rate": 0.00119886693594988, + "loss": 3.9373550415039062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.011331096291542053, + "learning_rate": 0.0011975646848367098, + "loss": 3.935594081878662, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.010538448579609394, + "learning_rate": 0.0011962626894754405, + "loss": 3.911579132080078, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.00989572424441576, + "learning_rate": 0.0011949609509150645, + "loss": 3.92525577545166, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.010533696971833706, + "learning_rate": 0.0011936594702043704, + "loss": 3.898075580596924, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.01051455456763506, + "learning_rate": 0.0011923582483919365, + "loss": 3.93463134765625, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.009662075899541378, + "learning_rate": 0.0011910572865261348, + "loss": 3.898221731185913, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.008874096907675266, + "learning_rate": 0.0011897565856551257, + "loss": 3.926787853240967, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.008126807399094105, + "learning_rate": 0.0011884561468268606, + "loss": 3.9158413410186768, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.007915553636848927, + "learning_rate": 0.001187155971089081, + "loss": 3.8891496658325195, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.008312922902405262, + "learning_rate": 0.0011858560594893129, + "loss": 3.932219982147217, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.008892328478395939, + "learning_rate": 0.0011845564130748734, + "loss": 3.8866336345672607, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.009187526069581509, + "learning_rate": 0.0011832570328928628, + "loss": 3.9464707374572754, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.008615494705736637, + "learning_rate": 0.0011819579199901693, + "loss": 3.9396629333496094, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.008649308234453201, + "learning_rate": 0.0011806590754134636, + "loss": 3.9325060844421387, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.008581082336604595, + "learning_rate": 0.0011793605002092021, + "loss": 3.9200944900512695, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.008502699434757233, + "learning_rate": 0.0011780621954236238, + "loss": 3.9656100273132324, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.0083576999604702, + "learning_rate": 0.0011767641621027477, + "loss": 3.9130215644836426, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.008196269161999226, + "learning_rate": 0.0011754664012923785, + "loss": 3.942369222640991, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.008857307955622673, + "learning_rate": 0.0011741689140380956, + "loss": 3.909585952758789, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.008935349993407726, + "learning_rate": 0.0011728717013852635, + "loss": 3.9248366355895996, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.008821200579404831, + "learning_rate": 0.0011715747643790211, + "loss": 3.9589288234710693, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.00861942209303379, + "learning_rate": 0.0011702781040642885, + "loss": 3.9381728172302246, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.009127574972808361, + "learning_rate": 0.0011689817214857613, + "loss": 3.897829532623291, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.008008269593119621, + "learning_rate": 0.0011676856176879107, + "loss": 3.930056095123291, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.010070208460092545, + "learning_rate": 0.0011663897937149853, + "loss": 3.892270803451538, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.00903113093227148, + "learning_rate": 0.001165094250611006, + "loss": 3.9766440391540527, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.008872415870428085, + "learning_rate": 0.0011637989894197692, + "loss": 3.9559712409973145, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.009524816647171974, + "learning_rate": 0.0011625040111848427, + "loss": 3.9607882499694824, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.009836779907345772, + "learning_rate": 0.0011612093169495678, + "loss": 3.9629592895507812, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.009586540050804615, + "learning_rate": 0.0011599149077570561, + "loss": 3.95208740234375, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.010495901107788086, + "learning_rate": 0.0011586207846501896, + "loss": 3.9439730644226074, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.010346266441047192, + "learning_rate": 0.0011573269486716202, + "loss": 3.959299087524414, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.010767634026706219, + "learning_rate": 0.0011560334008637674, + "loss": 3.900784969329834, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.012122200801968575, + "learning_rate": 0.0011547401422688202, + "loss": 3.9588193893432617, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.012914645485579967, + "learning_rate": 0.0011534471739287331, + "loss": 3.9278407096862793, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.011785333044826984, + "learning_rate": 0.0011521544968852273, + "loss": 3.992793560028076, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.008455781266093254, + "learning_rate": 0.0011508621121797899, + "loss": 3.96525239944458, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.009655282832682133, + "learning_rate": 0.0011495700208536714, + "loss": 3.9346890449523926, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.010537004098296165, + "learning_rate": 0.0011482782239478871, + "loss": 3.9534668922424316, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.009438651613891125, + "learning_rate": 0.0011469867225032135, + "loss": 3.983487606048584, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.010586646385490894, + "learning_rate": 0.0011456955175601908, + "loss": 3.939680576324463, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.009082718752324581, + "learning_rate": 0.001144404610159119, + "loss": 3.9558515548706055, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.009833602234721184, + "learning_rate": 0.0011431140013400591, + "loss": 3.9539828300476074, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.009142170660197735, + "learning_rate": 0.0011418236921428313, + "loss": 3.9442849159240723, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.008603853173553944, + "learning_rate": 0.0011405336836070137, + "loss": 3.9526913166046143, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.008838302455842495, + "learning_rate": 0.0011392439767719446, + "loss": 3.973104953765869, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.008867006748914719, + "learning_rate": 0.0011379545726767156, + "loss": 3.9832305908203125, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.00971697736531496, + "learning_rate": 0.001136665472360178, + "loss": 3.9563937187194824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.00957175437361002, + "learning_rate": 0.001135376676860935, + "loss": 3.975194215774536, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.008548660203814507, + "learning_rate": 0.001134088187217347, + "loss": 3.9839985370635986, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.008556334301829338, + "learning_rate": 0.0011328000044675264, + "loss": 3.9225516319274902, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.008440867997705936, + "learning_rate": 0.0011315121296493385, + "loss": 3.945028066635132, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.009643333032727242, + "learning_rate": 0.0011302245638004017, + "loss": 3.943120002746582, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.008832835592329502, + "learning_rate": 0.001128937307958083, + "loss": 3.9895920753479004, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.008720474317669868, + "learning_rate": 0.0011276503631595036, + "loss": 3.941950798034668, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.008099926635622978, + "learning_rate": 0.001126363730441529, + "loss": 3.9626870155334473, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.008525107055902481, + "learning_rate": 0.0011250774108407778, + "loss": 3.950535774230957, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.008998272009193897, + "learning_rate": 0.0011237914053936135, + "loss": 3.9460678100585938, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.009387437254190445, + "learning_rate": 0.001122505715136148, + "loss": 3.926604986190796, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.007804017513990402, + "learning_rate": 0.0011212203411042386, + "loss": 3.9894230365753174, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.008564616553485394, + "learning_rate": 0.0011199352843334876, + "loss": 3.9339959621429443, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.008726440370082855, + "learning_rate": 0.0011186505458592427, + "loss": 3.9867987632751465, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.008074100129306316, + "learning_rate": 0.0011173661267165932, + "loss": 3.9492292404174805, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.007688041776418686, + "learning_rate": 0.0011160820279403744, + "loss": 3.9425454139709473, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.007808993570506573, + "learning_rate": 0.0011147982505651595, + "loss": 3.935114860534668, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.009786820039153099, + "learning_rate": 0.0011135147956252659, + "loss": 3.959758758544922, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.009487155824899673, + "learning_rate": 0.00111223166415475, + "loss": 3.989048480987549, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.008407804183661938, + "learning_rate": 0.0011109488571874073, + "loss": 3.932347297668457, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.007158380467444658, + "learning_rate": 0.0011096663757567723, + "loss": 3.9545154571533203, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.008840581402182579, + "learning_rate": 0.001108384220896117, + "loss": 3.965984344482422, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.409194469451904, + "eval_runtime": 51.5561, + "eval_samples_per_second": 47.366, + "eval_steps_per_second": 1.494, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.010702896863222122, + "learning_rate": 0.0011071023936384508, + "loss": 3.8964462280273438, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01290619745850563, + "learning_rate": 0.0011058208950165183, + "loss": 3.8775625228881836, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01443667896091938, + "learning_rate": 0.0011045397260628, + "loss": 3.8755288124084473, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.014411968179047108, + "learning_rate": 0.0011032588878095108, + "loss": 3.889918327331543, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.013656266033649445, + "learning_rate": 0.0011019783812885986, + "loss": 3.926997661590576, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.010661937296390533, + "learning_rate": 0.0011006982075317446, + "loss": 3.9226956367492676, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.008968768641352654, + "learning_rate": 0.0010994183675703615, + "loss": 3.8932647705078125, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.009992181323468685, + "learning_rate": 0.0010981388624355937, + "loss": 3.8920068740844727, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.011034460738301277, + "learning_rate": 0.0010968596931583148, + "loss": 3.909073829650879, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.011815821751952171, + "learning_rate": 0.0010955808607691284, + "loss": 3.9190006256103516, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.012317490763962269, + "learning_rate": 0.0010943023662983674, + "loss": 3.8790552616119385, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.010983958840370178, + "learning_rate": 0.0010930242107760903, + "loss": 3.8730852603912354, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.008948802947998047, + "learning_rate": 0.0010917463952320856, + "loss": 3.923659563064575, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.009568693116307259, + "learning_rate": 0.0010904689206958649, + "loss": 3.9100794792175293, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.011862963438034058, + "learning_rate": 0.0010891917881966672, + "loss": 3.908093214035034, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.010180995799601078, + "learning_rate": 0.0010879149987634545, + "loss": 3.900660514831543, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.009372464381158352, + "learning_rate": 0.0010866385534249131, + "loss": 3.8950703144073486, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.010695629753172398, + "learning_rate": 0.0010853624532094525, + "loss": 3.913530111312866, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.011817459017038345, + "learning_rate": 0.0010840866991452024, + "loss": 3.90268611907959, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.010686337016522884, + "learning_rate": 0.0010828112922600167, + "loss": 3.9141364097595215, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.009550466202199459, + "learning_rate": 0.0010815362335814655, + "loss": 3.968468189239502, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.009367208927869797, + "learning_rate": 0.0010802615241368423, + "loss": 3.8933968544006348, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.009899397380650043, + "learning_rate": 0.0010789871649531562, + "loss": 3.936708450317383, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.009181005880236626, + "learning_rate": 0.0010777131570571363, + "loss": 3.9370503425598145, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.00799303874373436, + "learning_rate": 0.0010764395014752274, + "loss": 3.911050319671631, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.008469315245747566, + "learning_rate": 0.0010751661992335908, + "loss": 3.9103240966796875, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.009814459830522537, + "learning_rate": 0.0010738932513581034, + "loss": 3.881843090057373, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.010244108736515045, + "learning_rate": 0.0010726206588743557, + "loss": 3.883141040802002, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.008898732252418995, + "learning_rate": 0.001071348422807653, + "loss": 3.952347755432129, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.00823870487511158, + "learning_rate": 0.0010700765441830121, + "loss": 3.932507038116455, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.009857657365500927, + "learning_rate": 0.001068805024025164, + "loss": 3.932932138442993, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.00929308496415615, + "learning_rate": 0.0010675338633585488, + "loss": 3.8971338272094727, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.008667961694300175, + "learning_rate": 0.0010662630632073172, + "loss": 3.916931629180908, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.009758287109434605, + "learning_rate": 0.001064992624595331, + "loss": 3.964726209640503, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.009429254569113255, + "learning_rate": 0.0010637225485461586, + "loss": 3.917572498321533, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.008303353562951088, + "learning_rate": 0.0010624528360830782, + "loss": 3.875241756439209, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.008729013614356518, + "learning_rate": 0.001061183488229073, + "loss": 3.9691600799560547, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.010239302180707455, + "learning_rate": 0.0010599145060068353, + "loss": 3.9053149223327637, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.00932721421122551, + "learning_rate": 0.0010586458904387596, + "loss": 3.8593955039978027, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.008196988143026829, + "learning_rate": 0.0010573776425469473, + "loss": 3.917983055114746, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.008049141615629196, + "learning_rate": 0.001056109763353203, + "loss": 3.9337785243988037, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.009256313554942608, + "learning_rate": 0.001054842253879034, + "loss": 3.937887668609619, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.009292179718613625, + "learning_rate": 0.0010535751151456495, + "loss": 3.9203450679779053, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.009142161346971989, + "learning_rate": 0.0010523083481739602, + "loss": 3.9027457237243652, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.009498288854956627, + "learning_rate": 0.0010510419539845777, + "loss": 3.917316436767578, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.00849096104502678, + "learning_rate": 0.001049775933597813, + "loss": 3.9229235649108887, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.008343606255948544, + "learning_rate": 0.0010485102880336747, + "loss": 3.9138073921203613, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.010342134162783623, + "learning_rate": 0.0010472450183118727, + "loss": 3.8932929039001465, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.010298306122422218, + "learning_rate": 0.00104598012545181, + "loss": 3.9112677574157715, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.00827560480684042, + "learning_rate": 0.001044715610472589, + "loss": 3.894556760787964, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.008993102237582207, + "learning_rate": 0.0010434514743930064, + "loss": 3.905391216278076, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.008632538840174675, + "learning_rate": 0.0010421877182315536, + "loss": 3.896714210510254, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.008542424067854881, + "learning_rate": 0.0010409243430064162, + "loss": 3.9145569801330566, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.008970450609922409, + "learning_rate": 0.0010396613497354725, + "loss": 3.9170360565185547, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.008744142949581146, + "learning_rate": 0.0010383987394362942, + "loss": 3.927412509918213, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.008229847066104412, + "learning_rate": 0.0010371365131261424, + "loss": 3.928352117538452, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.00818693358451128, + "learning_rate": 0.0010358746718219717, + "loss": 3.9219343662261963, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.008848588913679123, + "learning_rate": 0.001034613216540423, + "loss": 3.9346699714660645, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.00834780652076006, + "learning_rate": 0.0010333521482978295, + "loss": 3.9144492149353027, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.00940527394413948, + "learning_rate": 0.0010320914681102098, + "loss": 3.951573371887207, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.008884521201252937, + "learning_rate": 0.0010308311769932725, + "loss": 3.899625778198242, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.007863261736929417, + "learning_rate": 0.0010295712759624107, + "loss": 3.9321727752685547, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.008651153184473515, + "learning_rate": 0.0010283117660327035, + "loss": 3.9030094146728516, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.00822074618190527, + "learning_rate": 0.0010270526482189163, + "loss": 3.939466953277588, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.008701153099536896, + "learning_rate": 0.0010257939235354964, + "loss": 3.9099626541137695, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.008418620564043522, + "learning_rate": 0.0010245355929965774, + "loss": 3.914116382598877, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.008339700289070606, + "learning_rate": 0.001023277657615971, + "loss": 3.9111878871917725, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.008304647170007229, + "learning_rate": 0.0010220201184071747, + "loss": 3.950974464416504, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.008436229079961777, + "learning_rate": 0.001020762976383365, + "loss": 3.9631457328796387, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.0074487063102424145, + "learning_rate": 0.001019506232557398, + "loss": 3.902984619140625, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.008736515417695045, + "learning_rate": 0.0010182498879418096, + "loss": 3.9268949031829834, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.009666468016803265, + "learning_rate": 0.0010169939435488145, + "loss": 3.9274301528930664, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.395970821380615, + "eval_runtime": 53.7848, + "eval_samples_per_second": 45.403, + "eval_steps_per_second": 1.432, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.009094545617699623, + "learning_rate": 0.001015738400390304, + "loss": 3.866425037384033, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.009273733012378216, + "learning_rate": 0.0010144832594778462, + "loss": 3.8569841384887695, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.009205271489918232, + "learning_rate": 0.0010132285218226855, + "loss": 3.8981378078460693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.009963045828044415, + "learning_rate": 0.0010119741884357424, + "loss": 3.881542444229126, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.008536554872989655, + "learning_rate": 0.0010107202603276096, + "loss": 3.851253032684326, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.01074561569839716, + "learning_rate": 0.001009466738508555, + "loss": 3.8731184005737305, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.010809902101755142, + "learning_rate": 0.0010082136239885177, + "loss": 3.863889217376709, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.008289352990686893, + "learning_rate": 0.0010069609177771103, + "loss": 3.839066743850708, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.00792208593338728, + "learning_rate": 0.001005708620883615, + "loss": 3.8634843826293945, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.008381598629057407, + "learning_rate": 0.0010044567343169845, + "loss": 3.868283987045288, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.009784617461264133, + "learning_rate": 0.0010032052590858425, + "loss": 3.853074550628662, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.009462198242545128, + "learning_rate": 0.0010019541961984782, + "loss": 3.889979600906372, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.010207819752395153, + "learning_rate": 0.001000703546662852, + "loss": 3.8636741638183594, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.009123551659286022, + "learning_rate": 0.0009994533114865887, + "loss": 3.8578555583953857, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.009006140753626823, + "learning_rate": 0.0009982034916769806, + "loss": 3.8828015327453613, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.0091114342212677, + "learning_rate": 0.0009969540882409846, + "loss": 3.880486011505127, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.008851367980241776, + "learning_rate": 0.0009957051021852227, + "loss": 3.8618125915527344, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.008196741342544556, + "learning_rate": 0.0009944565345159806, + "loss": 3.8314175605773926, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.00863513071089983, + "learning_rate": 0.000993208386239206, + "loss": 3.856212615966797, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.008624681271612644, + "learning_rate": 0.0009919606583605107, + "loss": 3.868927001953125, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.009208842180669308, + "learning_rate": 0.0009907133518851647, + "loss": 3.9072189331054688, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.010200913064181805, + "learning_rate": 0.0009894664678181027, + "loss": 3.875272750854492, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.009892710484564304, + "learning_rate": 0.000988220007163914, + "loss": 3.888686418533325, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.009873585775494576, + "learning_rate": 0.000986973970926851, + "loss": 3.8607118129730225, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.009322221390902996, + "learning_rate": 0.0009857283601108228, + "loss": 3.8852052688598633, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.008141528815031052, + "learning_rate": 0.0009844831757193947, + "loss": 3.888213872909546, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.008191613480448723, + "learning_rate": 0.00098323841875579, + "loss": 3.9073967933654785, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.008419081568717957, + "learning_rate": 0.0009819940902228864, + "loss": 3.879223346710205, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.009037869982421398, + "learning_rate": 0.0009807501911232171, + "loss": 3.883303165435791, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.00854078121483326, + "learning_rate": 0.0009795067224589687, + "loss": 3.926767587661743, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.008590771816670895, + "learning_rate": 0.0009782636852319825, + "loss": 3.913999557495117, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.00859815627336502, + "learning_rate": 0.0009770210804437498, + "loss": 3.8828063011169434, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.008970260620117188, + "learning_rate": 0.0009757789090954157, + "loss": 3.9174866676330566, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.00877175759524107, + "learning_rate": 0.0009745371721877754, + "loss": 3.8879401683807373, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.008541764691472054, + "learning_rate": 0.0009732958707212731, + "loss": 3.8717427253723145, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.008466241881251335, + "learning_rate": 0.0009720550056960039, + "loss": 3.8944485187530518, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.00808491837233305, + "learning_rate": 0.0009708145781117096, + "loss": 3.8617026805877686, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.00842207670211792, + "learning_rate": 0.0009695745889677805, + "loss": 3.887824535369873, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.00830599945038557, + "learning_rate": 0.0009683350392632533, + "loss": 3.8612070083618164, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.008417914621531963, + "learning_rate": 0.0009670959299968111, + "loss": 3.880140781402588, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.008542059920728207, + "learning_rate": 0.0009658572621667822, + "loss": 3.8951029777526855, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.008822236210107803, + "learning_rate": 0.000964619036771138, + "loss": 3.8962583541870117, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.009151184931397438, + "learning_rate": 0.0009633812548074952, + "loss": 3.882500648498535, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.007803036365658045, + "learning_rate": 0.0009621439172731115, + "loss": 3.8740782737731934, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.009350850246846676, + "learning_rate": 0.0009609070251648884, + "loss": 3.8627419471740723, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.008319144137203693, + "learning_rate": 0.0009596705794793667, + "loss": 3.9084486961364746, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.009198528714478016, + "learning_rate": 0.0009584345812127282, + "loss": 3.92141056060791, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.010480938479304314, + "learning_rate": 0.0009571990313607958, + "loss": 3.8916103839874268, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.01045288611203432, + "learning_rate": 0.0009559639309190278, + "loss": 3.8863210678100586, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.009121391922235489, + "learning_rate": 0.0009547292808825239, + "loss": 3.9167044162750244, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.01093223039060831, + "learning_rate": 0.0009534950822460184, + "loss": 3.9046711921691895, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.01214868575334549, + "learning_rate": 0.0009522613360038835, + "loss": 3.9126768112182617, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.01055280864238739, + "learning_rate": 0.0009510280431501255, + "loss": 3.9214353561401367, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.009021313861012459, + "learning_rate": 0.0009497952046783867, + "loss": 3.8795580863952637, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.00863752793520689, + "learning_rate": 0.0009485628215819428, + "loss": 3.893232822418213, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.009327626787126064, + "learning_rate": 0.0009473308948537021, + "loss": 3.917377471923828, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.01003898587077856, + "learning_rate": 0.0009460994254862066, + "loss": 3.8818931579589844, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.00892532430589199, + "learning_rate": 0.0009448684144716278, + "loss": 3.8761978149414062, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.008396463468670845, + "learning_rate": 0.0009436378628017702, + "loss": 3.8665928840637207, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.00784370768815279, + "learning_rate": 0.0009424077714680662, + "loss": 3.8850138187408447, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.00809321366250515, + "learning_rate": 0.0009411781414615784, + "loss": 3.917078733444214, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.008685288950800896, + "learning_rate": 0.000939948973772998, + "loss": 3.8865013122558594, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.007291820365935564, + "learning_rate": 0.0009387202693926425, + "loss": 3.8967666625976562, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.008459319360554218, + "learning_rate": 0.0009374920293104573, + "loss": 3.9043235778808594, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.007924137637019157, + "learning_rate": 0.0009362642545160123, + "loss": 3.9216856956481934, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.007874754257500172, + "learning_rate": 0.0009350369459985052, + "loss": 3.879725933074951, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.008643723092973232, + "learning_rate": 0.0009338101047467544, + "loss": 3.9303886890411377, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.008451704867184162, + "learning_rate": 0.000932583731749205, + "loss": 3.9352827072143555, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.008137754164636135, + "learning_rate": 0.0009313578279939236, + "loss": 3.930800437927246, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.009277567267417908, + "learning_rate": 0.0009301323944685979, + "loss": 3.9194300174713135, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.009428970515727997, + "learning_rate": 0.0009289074321605383, + "loss": 3.874298572540283, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.0089727733284235, + "learning_rate": 0.0009276829420566743, + "loss": 3.893770694732666, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.388528823852539, + "eval_runtime": 51.5859, + "eval_samples_per_second": 47.338, + "eval_steps_per_second": 1.493, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.009470812045037746, + "learning_rate": 0.000926458925143556, + "loss": 3.8514318466186523, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.011485518887639046, + "learning_rate": 0.0009252353824073511, + "loss": 3.857734203338623, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.010095939971506596, + "learning_rate": 0.0009240123148338457, + "loss": 3.819117546081543, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.010103728622198105, + "learning_rate": 0.0009227897234084444, + "loss": 3.842660427093506, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.01148147787898779, + "learning_rate": 0.0009215676091161661, + "loss": 3.821089029312134, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.012331553734838963, + "learning_rate": 0.0009203459729416467, + "loss": 3.8194339275360107, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.011274794116616249, + "learning_rate": 0.0009191248158691358, + "loss": 3.8839240074157715, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.009603365324437618, + "learning_rate": 0.0009179041388824986, + "loss": 3.806617259979248, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.010824795812368393, + "learning_rate": 0.0009166839429652112, + "loss": 3.823228597640991, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.01014280878007412, + "learning_rate": 0.0009154642291003637, + "loss": 3.860605239868164, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.009052124805748463, + "learning_rate": 0.0009142449982706591, + "loss": 3.84226131439209, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.009465130977332592, + "learning_rate": 0.0009130262514584074, + "loss": 3.8674912452697754, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.009991783648729324, + "learning_rate": 0.0009118079896455326, + "loss": 3.8775739669799805, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.009288253262639046, + "learning_rate": 0.0009105902138135652, + "loss": 3.7978098392486572, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.009964323602616787, + "learning_rate": 0.0009093729249436459, + "loss": 3.8362321853637695, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.010977157391607761, + "learning_rate": 0.0009081561240165218, + "loss": 3.857304096221924, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.009241319261491299, + "learning_rate": 0.0009069398120125479, + "loss": 3.8659310340881348, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.009211038239300251, + "learning_rate": 0.0009057239899116844, + "loss": 3.8764538764953613, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.009649050422012806, + "learning_rate": 0.000904508658693497, + "loss": 3.8702778816223145, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.00954380165785551, + "learning_rate": 0.0009032938193371573, + "loss": 3.8273327350616455, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.008831270970404148, + "learning_rate": 0.0009020794728214378, + "loss": 3.821664810180664, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.008367286063730717, + "learning_rate": 0.000900865620124717, + "loss": 3.8586034774780273, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.008805174380540848, + "learning_rate": 0.0008996522622249732, + "loss": 3.811522960662842, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.00873226672410965, + "learning_rate": 0.0008984394000997873, + "loss": 3.8538150787353516, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.008930203504860401, + "learning_rate": 0.0008972270347263402, + "loss": 3.8448486328125, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.009300434030592442, + "learning_rate": 0.0008960151670814125, + "loss": 3.842113494873047, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.009598663076758385, + "learning_rate": 0.0008948037981413855, + "loss": 3.886852741241455, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.008400370366871357, + "learning_rate": 0.0008935929288822357, + "loss": 3.852076292037964, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.00891344528645277, + "learning_rate": 0.0008923825602795391, + "loss": 3.8480606079101562, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.00869396049529314, + "learning_rate": 0.0008911726933084678, + "loss": 3.8164706230163574, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.007914266549050808, + "learning_rate": 0.0008899633289437904, + "loss": 3.8768272399902344, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.008398533798754215, + "learning_rate": 0.0008887544681598687, + "loss": 3.8625502586364746, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.00887146033346653, + "learning_rate": 0.0008875461119306601, + "loss": 3.8934593200683594, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.008642833679914474, + "learning_rate": 0.0008863382612297172, + "loss": 3.8391976356506348, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.008099788799881935, + "learning_rate": 0.0008851309170301816, + "loss": 3.827965497970581, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.008166342042386532, + "learning_rate": 0.0008839240803047894, + "loss": 3.8999152183532715, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.009095479734241962, + "learning_rate": 0.0008827177520258671, + "loss": 3.876927614212036, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.009246554225683212, + "learning_rate": 0.0008815119331653321, + "loss": 3.8692450523376465, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.007059223018586636, + "learning_rate": 0.0008803066246946898, + "loss": 3.869400978088379, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.008499744348227978, + "learning_rate": 0.0008791018275850363, + "loss": 3.8303189277648926, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.007667839527130127, + "learning_rate": 0.0008778975428070543, + "loss": 3.827998638153076, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.00765089550986886, + "learning_rate": 0.0008766937713310145, + "loss": 3.8554482460021973, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.00862938817590475, + "learning_rate": 0.0008754905141267749, + "loss": 3.905458927154541, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.008733712136745453, + "learning_rate": 0.0008742877721637758, + "loss": 3.8564133644104004, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.0071738120168447495, + "learning_rate": 0.0008730855464110471, + "loss": 3.878807544708252, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.008418241515755653, + "learning_rate": 0.0008718838378371989, + "loss": 3.8399720191955566, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.008526502177119255, + "learning_rate": 0.0008706826474104263, + "loss": 3.847656488418579, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.00732462527230382, + "learning_rate": 0.0008694819760985069, + "loss": 3.8839492797851562, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.008424196392297745, + "learning_rate": 0.0008682818248688003, + "loss": 3.861239433288574, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.008984955959022045, + "learning_rate": 0.0008670821946882465, + "loss": 3.856529712677002, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.009004034101963043, + "learning_rate": 0.000865883086523365, + "loss": 3.8514528274536133, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.009097260423004627, + "learning_rate": 0.0008646845013402563, + "loss": 3.8890581130981445, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.009244748391211033, + "learning_rate": 0.0008634864401045988, + "loss": 3.872745990753174, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.00788815412670374, + "learning_rate": 0.0008622889037816488, + "loss": 3.8777575492858887, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.008943610824644566, + "learning_rate": 0.0008610918933362396, + "loss": 3.8825736045837402, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.008154108189046383, + "learning_rate": 0.0008598954097327809, + "loss": 3.877514362335205, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.008689315989613533, + "learning_rate": 0.0008586994539352587, + "loss": 3.854720115661621, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.00904505793005228, + "learning_rate": 0.0008575040269072315, + "loss": 3.911463499069214, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.007773210760205984, + "learning_rate": 0.0008563091296118342, + "loss": 3.875370740890503, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.008437697775661945, + "learning_rate": 0.0008551147630117737, + "loss": 3.8457469940185547, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.009780957363545895, + "learning_rate": 0.0008539209280693296, + "loss": 3.890479564666748, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.007929213345050812, + "learning_rate": 0.0008527276257463532, + "loss": 3.876030445098877, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.009121386334300041, + "learning_rate": 0.0008515348570042669, + "loss": 3.9252734184265137, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.00981531199067831, + "learning_rate": 0.000850342622804063, + "loss": 3.8997983932495117, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.009150844067335129, + "learning_rate": 0.0008491509241063025, + "loss": 3.85115385055542, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.00992978923022747, + "learning_rate": 0.0008479597618711154, + "loss": 3.9138102531433105, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.010906689800322056, + "learning_rate": 0.0008467691370582001, + "loss": 3.892329692840576, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.008644519373774529, + "learning_rate": 0.0008455790506268213, + "loss": 3.9071311950683594, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.009030954912304878, + "learning_rate": 0.0008443895035358111, + "loss": 3.867384910583496, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.010806960985064507, + "learning_rate": 0.0008432004967435638, + "loss": 3.877983331680298, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.011148087680339813, + "learning_rate": 0.0008420120312080436, + "loss": 3.8795924186706543, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01046178862452507, + "learning_rate": 0.0008408241078867731, + "loss": 3.887709617614746, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.383835792541504, + "eval_runtime": 53.0778, + "eval_samples_per_second": 46.008, + "eval_steps_per_second": 1.451, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.009721867740154266, + "learning_rate": 0.0008396367277368421, + "loss": 3.8261327743530273, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.01091420091688633, + "learning_rate": 0.000838449891714901, + "loss": 3.7837276458740234, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.010423852130770683, + "learning_rate": 0.0008372636007771623, + "loss": 3.796450138092041, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.011928882449865341, + "learning_rate": 0.0008360778558793996, + "loss": 3.794717311859131, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.010779444128274918, + "learning_rate": 0.000834892657976945, + "loss": 3.8477983474731445, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.009554282762110233, + "learning_rate": 0.0008337080080246915, + "loss": 3.793790817260742, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.009768152609467506, + "learning_rate": 0.0008325239069770906, + "loss": 3.860384225845337, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.010801832191646099, + "learning_rate": 0.0008313403557881513, + "loss": 3.816796064376831, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.01046004518866539, + "learning_rate": 0.0008301573554114378, + "loss": 3.8064301013946533, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.009319447912275791, + "learning_rate": 0.0008289749068000744, + "loss": 3.832408905029297, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01011407095938921, + "learning_rate": 0.000827793010906737, + "loss": 3.8276009559631348, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.010393989272415638, + "learning_rate": 0.0008266116686836582, + "loss": 3.83754825592041, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.00883004255592823, + "learning_rate": 0.0008254308810826244, + "loss": 3.8315513134002686, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.008549787104129791, + "learning_rate": 0.0008242506490549745, + "loss": 3.792919158935547, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.009668390266597271, + "learning_rate": 0.0008230709735516006, + "loss": 3.8110954761505127, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.009008699096739292, + "learning_rate": 0.0008218918555229445, + "loss": 3.8598508834838867, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.008660725317895412, + "learning_rate": 0.0008207132959190026, + "loss": 3.839632511138916, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.00827972125262022, + "learning_rate": 0.0008195352956893172, + "loss": 3.8428683280944824, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.008117233403027058, + "learning_rate": 0.0008183578557829825, + "loss": 3.8434956073760986, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.008314455859363079, + "learning_rate": 0.0008171809771486405, + "loss": 3.8079686164855957, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.00865850131958723, + "learning_rate": 0.0008160046607344811, + "loss": 3.846999406814575, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0096864253282547, + "learning_rate": 0.0008148289074882416, + "loss": 3.8227243423461914, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.00840236246585846, + "learning_rate": 0.0008136537183572045, + "loss": 3.8400344848632812, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.007451791316270828, + "learning_rate": 0.0008124790942881984, + "loss": 3.8318095207214355, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.00858070608228445, + "learning_rate": 0.000811305036227597, + "loss": 3.8232922554016113, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.00765061192214489, + "learning_rate": 0.0008101315451213178, + "loss": 3.81624174118042, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.009304090403020382, + "learning_rate": 0.0008089586219148213, + "loss": 3.8715004920959473, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.009136030450463295, + "learning_rate": 0.0008077862675531103, + "loss": 3.825037956237793, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.007460939697921276, + "learning_rate": 0.0008066144829807303, + "loss": 3.811749219894409, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.007510480936616659, + "learning_rate": 0.0008054432691417659, + "loss": 3.8313679695129395, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.008825462311506271, + "learning_rate": 0.0008042726269798433, + "loss": 3.83119535446167, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.009002318605780602, + "learning_rate": 0.0008031025574381281, + "loss": 3.8542468547821045, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.009266122244298458, + "learning_rate": 0.0008019330614593237, + "loss": 3.8328065872192383, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.008935115300118923, + "learning_rate": 0.0008007641399856727, + "loss": 3.8198719024658203, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.008021850138902664, + "learning_rate": 0.0007995957939589526, + "loss": 3.828382968902588, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.008035171777009964, + "learning_rate": 0.0007984280243204805, + "loss": 3.879868507385254, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.008553579449653625, + "learning_rate": 0.0007972608320111059, + "loss": 3.842162609100342, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.008016746491193771, + "learning_rate": 0.000796094217971215, + "loss": 3.85674786567688, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.007781418971717358, + "learning_rate": 0.0007949281831407276, + "loss": 3.7968859672546387, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.008552255108952522, + "learning_rate": 0.0007937627284590971, + "loss": 3.8305506706237793, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.007390565704554319, + "learning_rate": 0.0007925978548653096, + "loss": 3.8228843212127686, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.007611850742250681, + "learning_rate": 0.0007914335632978812, + "loss": 3.868957996368408, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.008188549429178238, + "learning_rate": 0.0007902698546948628, + "loss": 3.836078405380249, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.008198797702789307, + "learning_rate": 0.0007891067299938318, + "loss": 3.8637044429779053, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.008353965356945992, + "learning_rate": 0.0007879441901318977, + "loss": 3.833843946456909, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.0086967209354043, + "learning_rate": 0.0007867822360456963, + "loss": 3.850536823272705, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.009151408448815346, + "learning_rate": 0.0007856208686713949, + "loss": 3.8453407287597656, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.008010942488908768, + "learning_rate": 0.0007844600889446858, + "loss": 3.8679513931274414, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.007813206873834133, + "learning_rate": 0.0007832998978007873, + "loss": 3.860609531402588, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.007759213913232088, + "learning_rate": 0.0007821402961744448, + "loss": 3.803863048553467, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.008045864291489124, + "learning_rate": 0.0007809812849999293, + "loss": 3.858161449432373, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.008212368935346603, + "learning_rate": 0.0007798228652110347, + "loss": 3.8311336040496826, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.007982932962477207, + "learning_rate": 0.0007786650377410774, + "loss": 3.870339870452881, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.0071402136236429214, + "learning_rate": 0.0007775078035229002, + "loss": 3.7922263145446777, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.008564750663936138, + "learning_rate": 0.0007763511634888656, + "loss": 3.83981990814209, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.008078585378825665, + "learning_rate": 0.0007751951185708562, + "loss": 3.852938652038574, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.00801754929125309, + "learning_rate": 0.0007740396697002774, + "loss": 3.8434009552001953, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.008973612450063229, + "learning_rate": 0.0007728848178080535, + "loss": 3.8238563537597656, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.010199865326285362, + "learning_rate": 0.0007717305638246281, + "loss": 3.8891735076904297, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.009620471857488155, + "learning_rate": 0.0007705769086799617, + "loss": 3.8611769676208496, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.007981563918292522, + "learning_rate": 0.0007694238533035337, + "loss": 3.8496928215026855, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.008856567554175854, + "learning_rate": 0.0007682713986243415, + "loss": 3.8366098403930664, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.008323253132402897, + "learning_rate": 0.0007671195455708952, + "loss": 3.8224244117736816, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.00814893003553152, + "learning_rate": 0.0007659682950712228, + "loss": 3.8611154556274414, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.009146672673523426, + "learning_rate": 0.0007648176480528659, + "loss": 3.8258843421936035, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.007573423441499472, + "learning_rate": 0.0007636676054428805, + "loss": 3.888007640838623, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.00813984964042902, + "learning_rate": 0.0007625181681678343, + "loss": 3.860306739807129, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.007936778478324413, + "learning_rate": 0.0007613693371538078, + "loss": 3.8808555603027344, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.008434669114649296, + "learning_rate": 0.0007602211133263956, + "loss": 3.854154348373413, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.007857564836740494, + "learning_rate": 0.0007590734976106985, + "loss": 3.8240737915039062, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.007989327423274517, + "learning_rate": 0.0007579264909313315, + "loss": 3.850971221923828, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.008888393640518188, + "learning_rate": 0.0007567800942124152, + "loss": 3.862267255783081, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.38022518157959, + "eval_runtime": 51.4686, + "eval_samples_per_second": 47.446, + "eval_steps_per_second": 1.496, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.00952722318470478, + "learning_rate": 0.0007556343083775832, + "loss": 3.7705495357513428, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01099323108792305, + "learning_rate": 0.0007544891343499723, + "loss": 3.80812406539917, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.010298571549355984, + "learning_rate": 0.0007533445730522295, + "loss": 3.782794237136841, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.008139589801430702, + "learning_rate": 0.000752200625406507, + "loss": 3.7877392768859863, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.009031252935528755, + "learning_rate": 0.0007510572923344628, + "loss": 3.8181564807891846, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.00942459236830473, + "learning_rate": 0.0007499145747572603, + "loss": 3.807425022125244, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.00922947097569704, + "learning_rate": 0.0007487724735955648, + "loss": 3.7923264503479004, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.009551790542900562, + "learning_rate": 0.0007476309897695485, + "loss": 3.790083885192871, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.008996761403977871, + "learning_rate": 0.0007464901241988831, + "loss": 3.836491107940674, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.009766182862222195, + "learning_rate": 0.0007453498778027433, + "loss": 3.7829647064208984, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.007708575110882521, + "learning_rate": 0.0007442102514998058, + "loss": 3.803400993347168, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.008606902323663235, + "learning_rate": 0.0007430712462082464, + "loss": 3.810825824737549, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.009462927468121052, + "learning_rate": 0.0007419328628457416, + "loss": 3.784748077392578, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.010283366777002811, + "learning_rate": 0.0007407951023294655, + "loss": 3.744312286376953, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.008952794596552849, + "learning_rate": 0.0007396579655760917, + "loss": 3.8098015785217285, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.008446207270026207, + "learning_rate": 0.0007385214535017907, + "loss": 3.8306760787963867, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.007988285273313522, + "learning_rate": 0.0007373855670222302, + "loss": 3.7976696491241455, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.007586437277495861, + "learning_rate": 0.0007362503070525718, + "loss": 3.785891056060791, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.00775605533272028, + "learning_rate": 0.0007351156745074757, + "loss": 3.8176379203796387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.007727691903710365, + "learning_rate": 0.0007339816703010948, + "loss": 3.791743278503418, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.008311517536640167, + "learning_rate": 0.0007328482953470747, + "loss": 3.818769693374634, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.007610874716192484, + "learning_rate": 0.0007317155505585556, + "loss": 3.76772141456604, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.0078107016161084175, + "learning_rate": 0.0007305834368481699, + "loss": 3.822758197784424, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.006906864698976278, + "learning_rate": 0.0007294519551280412, + "loss": 3.832834005355835, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.008230660110712051, + "learning_rate": 0.0007283211063097829, + "loss": 3.85087513923645, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.008201118558645248, + "learning_rate": 0.0007271908913044997, + "loss": 3.7777247428894043, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.0070146857760846615, + "learning_rate": 0.000726061311022787, + "loss": 3.837517738342285, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.0076036700047552586, + "learning_rate": 0.0007249323663747254, + "loss": 3.8358592987060547, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.007238952908664942, + "learning_rate": 0.0007238040582698858, + "loss": 3.828089952468872, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.00825269054621458, + "learning_rate": 0.0007226763876173259, + "loss": 3.806016445159912, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.008368568494915962, + "learning_rate": 0.0007215493553255898, + "loss": 3.8163394927978516, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.00728471577167511, + "learning_rate": 0.0007204229623027063, + "loss": 3.835597038269043, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.00784830842167139, + "learning_rate": 0.0007192972094561897, + "loss": 3.8050971031188965, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.007714141625910997, + "learning_rate": 0.0007181720976930404, + "loss": 3.8031606674194336, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.009128103032708168, + "learning_rate": 0.000717047627919739, + "loss": 3.77107572555542, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.008033272810280323, + "learning_rate": 0.000715923801042251, + "loss": 3.8615059852600098, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.007732565049082041, + "learning_rate": 0.0007148006179660235, + "loss": 3.8277132511138916, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.007382966112345457, + "learning_rate": 0.0007136780795959853, + "loss": 3.803431987762451, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.007517020218074322, + "learning_rate": 0.0007125561868365442, + "loss": 3.799734354019165, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.008041434921324253, + "learning_rate": 0.0007114349405915895, + "loss": 3.7848258018493652, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.007875649258494377, + "learning_rate": 0.0007103143417644891, + "loss": 3.8136563301086426, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.008213640190660954, + "learning_rate": 0.000709194391258089, + "loss": 3.8518495559692383, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.00801054760813713, + "learning_rate": 0.0007080750899747139, + "loss": 3.7956271171569824, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.008064007386565208, + "learning_rate": 0.0007069564388161631, + "loss": 3.854595184326172, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.008300521411001682, + "learning_rate": 0.0007058384386837154, + "loss": 3.8203625679016113, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.008640496991574764, + "learning_rate": 0.000704721090478122, + "loss": 3.793574810028076, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.007684200536459684, + "learning_rate": 0.0007036043950996105, + "loss": 3.8564438819885254, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.008431905880570412, + "learning_rate": 0.0007024883534478823, + "loss": 3.8271119594573975, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.008137186989188194, + "learning_rate": 0.0007013729664221122, + "loss": 3.8219337463378906, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.0076744635589420795, + "learning_rate": 0.0007002582349209474, + "loss": 3.8353049755096436, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.008388740941882133, + "learning_rate": 0.0006991441598425065, + "loss": 3.795104503631592, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.007856911979615688, + "learning_rate": 0.0006980307420843798, + "loss": 3.818411111831665, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.008048887364566326, + "learning_rate": 0.000696917982543628, + "loss": 3.86599063873291, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.007949979975819588, + "learning_rate": 0.0006958058821167812, + "loss": 3.8026742935180664, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.008162821643054485, + "learning_rate": 0.0006946944416998389, + "loss": 3.833751678466797, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.008190765976905823, + "learning_rate": 0.0006935836621882682, + "loss": 3.8164236545562744, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.009714310988783836, + "learning_rate": 0.000692473544477005, + "loss": 3.8367366790771484, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.008075987920165062, + "learning_rate": 0.0006913640894604498, + "loss": 3.8108773231506348, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.009075167588889599, + "learning_rate": 0.0006902552980324713, + "loss": 3.8444485664367676, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.008678440935909748, + "learning_rate": 0.0006891471710864022, + "loss": 3.8384461402893066, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.007705411408096552, + "learning_rate": 0.0006880397095150414, + "loss": 3.825443744659424, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.00782099086791277, + "learning_rate": 0.0006869329142106488, + "loss": 3.8331687450408936, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.007542501203715801, + "learning_rate": 0.0006858267860649509, + "loss": 3.7905991077423096, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.008315361104905605, + "learning_rate": 0.0006847213259691356, + "loss": 3.861527442932129, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.008529976941645145, + "learning_rate": 0.0006836165348138508, + "loss": 3.8332972526550293, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.008117211051285267, + "learning_rate": 0.0006825124134892072, + "loss": 3.7961089611053467, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.008409497328102589, + "learning_rate": 0.0006814089628847755, + "loss": 3.812218189239502, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.007592010777443647, + "learning_rate": 0.0006803061838895864, + "loss": 3.801304340362549, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.007820053026080132, + "learning_rate": 0.0006792040773921283, + "loss": 3.8517117500305176, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.0073044621385633945, + "learning_rate": 0.0006781026442803479, + "loss": 3.8253560066223145, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.007546546868979931, + "learning_rate": 0.0006770018854416522, + "loss": 3.800802230834961, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.008878474123775959, + "learning_rate": 0.0006759018017629007, + "loss": 3.8133859634399414, + "step": 2520 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.0743646764662784e+18, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-2520/training_args.bin b/runs/baseline/checkpoint-2520/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-2520/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-2880/chat_template.jinja b/runs/baseline/checkpoint-2880/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-2880/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-2880/config.json b/runs/baseline/checkpoint-2880/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-2880/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-2880/generation_config.json b/runs/baseline/checkpoint-2880/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-2880/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-2880/model.safetensors b/runs/baseline/checkpoint-2880/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33f73d5e9c8deebc108bef385e5e74b1ef8c8542 --- /dev/null +++ b/runs/baseline/checkpoint-2880/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:721348485e694b0135ed2c93c3d0877d9feeaaee742aa29f66371f9eba4ec67c +size 583356232 diff --git a/runs/baseline/checkpoint-2880/optimizer.pt b/runs/baseline/checkpoint-2880/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..9f30555c64a5f60c1d47dcc457c7bdce06ac3611 --- /dev/null +++ b/runs/baseline/checkpoint-2880/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:820cb1774150899abf0c334ccf7b12827b985cc98b69314900a502c06630797e +size 1166825803 diff --git a/runs/baseline/checkpoint-2880/rng_state_0.pth b/runs/baseline/checkpoint-2880/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..b73f9a380e4e9c61520ae0674c9453be2b007f8d --- /dev/null +++ b/runs/baseline/checkpoint-2880/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19e0f78f2d9435829a3e127bfda2ca14f3245c5620079efefde656536bbe279e +size 14917 diff --git a/runs/baseline/checkpoint-2880/rng_state_1.pth b/runs/baseline/checkpoint-2880/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..1c80ff79c801a37a1192dfdb25ad4404ef9b73aa --- /dev/null +++ b/runs/baseline/checkpoint-2880/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ff567f23d2e2493cacddea05b5da8eee7e39fe5f45b9a2a98aa6a743f71addc +size 14917 diff --git a/runs/baseline/checkpoint-2880/scheduler.pt b/runs/baseline/checkpoint-2880/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..fe8f25d9d984b4f8840e0a426b55887e7fc8a096 --- /dev/null +++ b/runs/baseline/checkpoint-2880/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c98bbcd7f4a448da5d1b7f87739b30169f3b0d703d298db87b75054e6b96f54a +size 1465 diff --git a/runs/baseline/checkpoint-2880/tokenizer.json b/runs/baseline/checkpoint-2880/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-2880/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-2880/tokenizer_config.json b/runs/baseline/checkpoint-2880/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-2880/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-2880/trainer_state.json b/runs/baseline/checkpoint-2880/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..461d29f690c3fe5229d1fba4311e1b133e9ad6dc --- /dev/null +++ b/runs/baseline/checkpoint-2880/trainer_state.json @@ -0,0 +1,20506 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 40.0, + "eval_steps": 500, + "global_step": 2880, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.641542434692383, + "eval_runtime": 51.6342, + "eval_samples_per_second": 47.294, + "eval_steps_per_second": 1.491, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.009694516658782959, + "learning_rate": 0.0020492498088767836, + "loss": 4.32808256149292, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.0097952950745821, + "learning_rate": 0.002048005909777114, + "loss": 4.280096054077148, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.010393359698355198, + "learning_rate": 0.0020467615812442106, + "loss": 4.317824363708496, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.00977372657507658, + "learning_rate": 0.0020455168242806054, + "loss": 4.3445329666137695, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.008568719029426575, + "learning_rate": 0.0020442716398891775, + "loss": 4.318026542663574, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.009732510894536972, + "learning_rate": 0.0020430260290731487, + "loss": 4.330609321594238, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.010429969057440758, + "learning_rate": 0.0020417799928360864, + "loss": 4.316794395446777, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.009292383678257465, + "learning_rate": 0.002040533532181898, + "loss": 4.317420959472656, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.011222455650568008, + "learning_rate": 0.0020392866481148356, + "loss": 4.292610168457031, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.013591725379228592, + "learning_rate": 0.0020380393416394894, + "loss": 4.325689315795898, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.013572026044130325, + "learning_rate": 0.002036791613760794, + "loss": 4.334960460662842, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.012986521236598492, + "learning_rate": 0.0020355434654840197, + "loss": 4.33829402923584, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.013321424834430218, + "learning_rate": 0.0020342948978147777, + "loss": 4.2926926612854, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.012307215481996536, + "learning_rate": 0.0020330459117590153, + "loss": 4.32757568359375, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.012909360229969025, + "learning_rate": 0.0020317965083230197, + "loss": 4.362889289855957, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017746610566973686, + "learning_rate": 0.0020305466885134115, + "loss": 4.3426313400268555, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.01696564070880413, + "learning_rate": 0.002029296453337148, + "loss": 4.3298516273498535, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.013348255306482315, + "learning_rate": 0.002028045803801522, + "loss": 4.360220432281494, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.013225145637989044, + "learning_rate": 0.002026794740914158, + "loss": 4.346250057220459, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.01569020375609398, + "learning_rate": 0.002025543265683016, + "loss": 4.308025360107422, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.018612321466207504, + "learning_rate": 0.002024291379116385, + "loss": 4.330190181732178, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.01619943603873253, + "learning_rate": 0.00202303908222289, + "loss": 4.339552402496338, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.015385213308036327, + "learning_rate": 0.0020217863760114826, + "loss": 4.3412184715271, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.018517112359404564, + "learning_rate": 0.002020533261491446, + "loss": 4.3289313316345215, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.014868509024381638, + "learning_rate": 0.0020192797396723903, + "loss": 4.359846115112305, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.014240248128771782, + "learning_rate": 0.0020180258115642578, + "loss": 4.369707107543945, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.018113158643245697, + "learning_rate": 0.0020167714781773143, + "loss": 4.362668514251709, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.01802460104227066, + "learning_rate": 0.002015516740522154, + "loss": 4.344742298126221, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.01414549071341753, + "learning_rate": 0.0020142615996096964, + "loss": 4.358080863952637, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.012157700955867767, + "learning_rate": 0.002013006056451186, + "loss": 4.347763538360596, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.011268534697592258, + "learning_rate": 0.0020117501120581905, + "loss": 4.335859298706055, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.010395628400146961, + "learning_rate": 0.002010493767442602, + "loss": 4.341609001159668, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.012147950939834118, + "learning_rate": 0.002009237023616635, + "loss": 4.328502655029297, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.00945813488215208, + "learning_rate": 0.0020079798815928252, + "loss": 4.311186790466309, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.009489500895142555, + "learning_rate": 0.0020067223423840294, + "loss": 4.371470928192139, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.009360029362142086, + "learning_rate": 0.0020054644070034234, + "loss": 4.33941125869751, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.010032779537141323, + "learning_rate": 0.0020042060764645037, + "loss": 4.365797519683838, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.009367046877741814, + "learning_rate": 0.002002947351781084, + "loss": 4.29443359375, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.008369733579456806, + "learning_rate": 0.0020016882339672967, + "loss": 4.294233322143555, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.010394785553216934, + "learning_rate": 0.0020004287240375893, + "loss": 4.352355003356934, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01068816613405943, + "learning_rate": 0.001999168823006728, + "loss": 4.342516899108887, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.009975973516702652, + "learning_rate": 0.0019979085318897905, + "loss": 4.346763610839844, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.009618151932954788, + "learning_rate": 0.001996647851702171, + "loss": 4.329115867614746, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.008399694226682186, + "learning_rate": 0.001995386783459577, + "loss": 4.28736686706543, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.008235268294811249, + "learning_rate": 0.001994125328178029, + "loss": 4.339872360229492, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.00836547277867794, + "learning_rate": 0.0019928634868738577, + "loss": 4.328156471252441, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.009619626216590405, + "learning_rate": 0.001991601260563706, + "loss": 4.337394714355469, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.009814038872718811, + "learning_rate": 0.0019903386502645275, + "loss": 4.325926780700684, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.009588954970240593, + "learning_rate": 0.001989075656993584, + "loss": 4.286477088928223, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.010471943765878677, + "learning_rate": 0.001987812281768447, + "loss": 4.322466850280762, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.010239035822451115, + "learning_rate": 0.001986548525606994, + "loss": 4.3487653732299805, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.008935568854212761, + "learning_rate": 0.001985284389527411, + "loss": 4.303601264953613, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.008624759502708912, + "learning_rate": 0.00198401987454819, + "loss": 4.34036922454834, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.009296547621488571, + "learning_rate": 0.0019827549816881277, + "loss": 4.273357391357422, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.009071138687431812, + "learning_rate": 0.0019814897119663248, + "loss": 4.321051597595215, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.008341389708220959, + "learning_rate": 0.0019802240664021876, + "loss": 4.343242645263672, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.007881715893745422, + "learning_rate": 0.0019789580460154227, + "loss": 4.32067346572876, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.008750553242862225, + "learning_rate": 0.0019776916518260404, + "loss": 4.330263614654541, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.008591460064053535, + "learning_rate": 0.001976424884854351, + "loss": 4.311957359313965, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.008084543980658054, + "learning_rate": 0.001975157746120966, + "loss": 4.331319808959961, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.008036594837903976, + "learning_rate": 0.001973890236646797, + "loss": 4.382656097412109, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.00805772002786398, + "learning_rate": 0.0019726223574530527, + "loss": 4.297247886657715, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.009019420482218266, + "learning_rate": 0.0019713541095612405, + "loss": 4.326526165008545, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.009953883476555347, + "learning_rate": 0.0019700854939931653, + "loss": 4.327356338500977, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.009340492077171803, + "learning_rate": 0.0019688165117709273, + "loss": 4.343993186950684, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.010395721532404423, + "learning_rate": 0.0019675471639169226, + "loss": 4.284934997558594, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.009946230798959732, + "learning_rate": 0.0019662774514538415, + "loss": 4.288936614990234, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.010981100611388683, + "learning_rate": 0.0019650073754046695, + "loss": 4.339935779571533, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.009688783437013626, + "learning_rate": 0.0019637369367926827, + "loss": 4.381425857543945, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.009442255832254887, + "learning_rate": 0.001962466136641452, + "loss": 4.304185390472412, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.009371291846036911, + "learning_rate": 0.0019611949759748364, + "loss": 4.306120872497559, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.009544814005494118, + "learning_rate": 0.001959923455816988, + "loss": 4.326251983642578, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.602745532989502, + "eval_runtime": 53.5041, + "eval_samples_per_second": 45.641, + "eval_steps_per_second": 1.439, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.01010364294052124, + "learning_rate": 0.0019586515771923474, + "loss": 4.24491024017334, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.012722729705274105, + "learning_rate": 0.0019573793411256446, + "loss": 4.257739543914795, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.01307204831391573, + "learning_rate": 0.0019561067486418967, + "loss": 4.291684150695801, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.011215966194868088, + "learning_rate": 0.0019548338007664098, + "loss": 4.237947463989258, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.012943019159138203, + "learning_rate": 0.0019535604985247727, + "loss": 4.242859363555908, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.01337282732129097, + "learning_rate": 0.0019522868429428636, + "loss": 4.238934516906738, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.013697991147637367, + "learning_rate": 0.0019510128350468438, + "loss": 4.285760402679443, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.012182159349322319, + "learning_rate": 0.001949738475863158, + "loss": 4.290044784545898, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.012620339170098305, + "learning_rate": 0.0019484637664185346, + "loss": 4.267187118530273, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.011284958571195602, + "learning_rate": 0.0019471887077399837, + "loss": 4.26662540435791, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.01161450706422329, + "learning_rate": 0.001945913300854797, + "loss": 4.283192157745361, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.010022968985140324, + "learning_rate": 0.001944637546790548, + "loss": 4.295843124389648, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.010341878980398178, + "learning_rate": 0.001943361446575087, + "loss": 4.295636177062988, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.011549185030162334, + "learning_rate": 0.0019420850012365457, + "loss": 4.288768768310547, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.012518810108304024, + "learning_rate": 0.0019408082118033331, + "loss": 4.229338645935059, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.012035328894853592, + "learning_rate": 0.0019395310793041353, + "loss": 4.258639335632324, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.011193782091140747, + "learning_rate": 0.0019382536047679146, + "loss": 4.239264488220215, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.009930136613547802, + "learning_rate": 0.0019369757892239096, + "loss": 4.252286434173584, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.00953229982405901, + "learning_rate": 0.0019356976337016327, + "loss": 4.293435096740723, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.010618340224027634, + "learning_rate": 0.0019344191392308716, + "loss": 4.235588550567627, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.008806651458144188, + "learning_rate": 0.0019331403068416856, + "loss": 4.26716423034668, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.008934796787798405, + "learning_rate": 0.0019318611375644065, + "loss": 4.258462905883789, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.009283283725380898, + "learning_rate": 0.0019305816324296386, + "loss": 4.257868766784668, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.009224812500178814, + "learning_rate": 0.0019293017924682555, + "loss": 4.258486747741699, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.008957969956099987, + "learning_rate": 0.0019280216187114016, + "loss": 4.309563636779785, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.010284051299095154, + "learning_rate": 0.0019267411121904894, + "loss": 4.275271415710449, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.009167762473225594, + "learning_rate": 0.0019254602739372001, + "loss": 4.238452911376953, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.009411636739969254, + "learning_rate": 0.0019241791049834816, + "loss": 4.2735595703125, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.009044171310961246, + "learning_rate": 0.0019228976063615497, + "loss": 4.27017879486084, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.008840068243443966, + "learning_rate": 0.0019216157791038834, + "loss": 4.28773307800293, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.009512359276413918, + "learning_rate": 0.001920333624243228, + "loss": 4.266008377075195, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.009946667589247227, + "learning_rate": 0.001919051142812593, + "loss": 4.279062271118164, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.00899894442409277, + "learning_rate": 0.0019177683358452505, + "loss": 4.277679443359375, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.00909668579697609, + "learning_rate": 0.0019164852043747338, + "loss": 4.311397552490234, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.009787327609956264, + "learning_rate": 0.0019152017494348408, + "loss": 4.30875825881958, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.008983899839222431, + "learning_rate": 0.001913917972059626, + "loss": 4.259649276733398, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.008411582559347153, + "learning_rate": 0.0019126338732834071, + "loss": 4.259811878204346, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.009162918664515018, + "learning_rate": 0.0019113494541407579, + "loss": 4.271515846252441, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.011304561980068684, + "learning_rate": 0.0019100647156665125, + "loss": 4.269217491149902, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.012483394704759121, + "learning_rate": 0.0019087796588957617, + "loss": 4.2949299812316895, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.012091380544006824, + "learning_rate": 0.0019074942848638522, + "loss": 4.289752006530762, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.010011567734181881, + "learning_rate": 0.0019062085946063869, + "loss": 4.294095993041992, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.008766217157244682, + "learning_rate": 0.0019049225891592226, + "loss": 4.256425857543945, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.012862488627433777, + "learning_rate": 0.001903636269558471, + "loss": 4.287847995758057, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.013917063362896442, + "learning_rate": 0.0019023496368404972, + "loss": 4.282784938812256, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.012129023671150208, + "learning_rate": 0.0019010626920419173, + "loss": 4.2486419677734375, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.010348426178097725, + "learning_rate": 0.0018997754361995986, + "loss": 4.289836883544922, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.011821693740785122, + "learning_rate": 0.001898487870350662, + "loss": 4.214344024658203, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.013136724010109901, + "learning_rate": 0.0018971999955324742, + "loss": 4.24208402633667, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.013173168525099754, + "learning_rate": 0.0018959118127826538, + "loss": 4.270164489746094, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.011018159799277782, + "learning_rate": 0.0018946233231390656, + "loss": 4.2571916580200195, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.01018658559769392, + "learning_rate": 0.0018933345276398227, + "loss": 4.283267498016357, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.009412688203155994, + "learning_rate": 0.0018920454273232848, + "loss": 4.272198677062988, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.007799267303198576, + "learning_rate": 0.0018907560232280557, + "loss": 4.260125637054443, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.008747464045882225, + "learning_rate": 0.0018894663163929863, + "loss": 4.2539215087890625, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.008003434166312218, + "learning_rate": 0.0018881763078571689, + "loss": 4.245024681091309, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.007593846879899502, + "learning_rate": 0.0018868859986599415, + "loss": 4.237929344177246, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.008099420927464962, + "learning_rate": 0.0018855953898408815, + "loss": 4.2789106369018555, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.008270486257970333, + "learning_rate": 0.0018843044824398097, + "loss": 4.263522148132324, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.009260379709303379, + "learning_rate": 0.0018830132774967866, + "loss": 4.275835037231445, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.009289492852985859, + "learning_rate": 0.0018817217760521132, + "loss": 4.313778400421143, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.009493921883404255, + "learning_rate": 0.0018804299791463281, + "loss": 4.303672790527344, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.009020023047924042, + "learning_rate": 0.0018791378878202105, + "loss": 4.245473384857178, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.009446556679904461, + "learning_rate": 0.0018778455031147724, + "loss": 4.267327785491943, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.009499974548816681, + "learning_rate": 0.0018765528260712672, + "loss": 4.301205635070801, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.008399393409490585, + "learning_rate": 0.00187525985773118, + "loss": 4.307802677154541, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.009519701823592186, + "learning_rate": 0.0018739665991362327, + "loss": 4.284796714782715, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01125232595950365, + "learning_rate": 0.0018726730513283804, + "loss": 4.285930633544922, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.011608537286520004, + "learning_rate": 0.00187137921534981, + "loss": 4.257689476013184, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.01064533181488514, + "learning_rate": 0.001870085092242944, + "loss": 4.249063968658447, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.009854599833488464, + "learning_rate": 0.0018687906830504321, + "loss": 4.255159378051758, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.01121739111840725, + "learning_rate": 0.0018674959888151576, + "loss": 4.287777900695801, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.5675368309021, + "eval_runtime": 51.8555, + "eval_samples_per_second": 47.092, + "eval_steps_per_second": 1.485, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.0104464590549469, + "learning_rate": 0.0018662010105802314, + "loss": 4.206080436706543, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.012800181284546852, + "learning_rate": 0.0018649057493889945, + "loss": 4.213550567626953, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.013382901437580585, + "learning_rate": 0.001863610206285015, + "loss": 4.190858840942383, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.014347050338983536, + "learning_rate": 0.0018623143823120894, + "loss": 4.22884464263916, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.01347152329981327, + "learning_rate": 0.001861018278514239, + "loss": 4.215579032897949, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.012786377221345901, + "learning_rate": 0.0018597218959357112, + "loss": 4.174999237060547, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.014137202873826027, + "learning_rate": 0.001858425235620979, + "loss": 4.21709680557251, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.013864399865269661, + "learning_rate": 0.001857128298614737, + "loss": 4.2375030517578125, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.010397658683359623, + "learning_rate": 0.0018558310859619048, + "loss": 4.231938362121582, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.00888819433748722, + "learning_rate": 0.0018545335987076223, + "loss": 4.195562839508057, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.010698067955672741, + "learning_rate": 0.001853235837897252, + "loss": 4.22847318649292, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.00987589918076992, + "learning_rate": 0.0018519378045763766, + "loss": 4.2147040367126465, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.011451782658696175, + "learning_rate": 0.001850639499790798, + "loss": 4.179263114929199, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.011224704794585705, + "learning_rate": 0.0018493409245865368, + "loss": 4.23030948638916, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.00996861606836319, + "learning_rate": 0.0018480420800098313, + "loss": 4.194283962249756, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.012143119238317013, + "learning_rate": 0.0018467429671071374, + "loss": 4.221704959869385, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.013126783072948456, + "learning_rate": 0.0018454435869251268, + "loss": 4.187748908996582, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.013031953014433384, + "learning_rate": 0.001844143940510687, + "loss": 4.248250961303711, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.012868993915617466, + "learning_rate": 0.001842844028910919, + "loss": 4.20161247253418, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.010102041065692902, + "learning_rate": 0.0018415438531731395, + "loss": 4.213654518127441, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.009599820710718632, + "learning_rate": 0.0018402434143448747, + "loss": 4.206313133239746, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.011536866426467896, + "learning_rate": 0.0018389427134738658, + "loss": 4.237072944641113, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.009558089077472687, + "learning_rate": 0.0018376417516080636, + "loss": 4.182195663452148, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01056219544261694, + "learning_rate": 0.0018363405297956302, + "loss": 4.232585906982422, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.012673584744334221, + "learning_rate": 0.0018350390490849354, + "loss": 4.228830337524414, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.011215884238481522, + "learning_rate": 0.0018337373105245598, + "loss": 4.216056823730469, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.011281915940344334, + "learning_rate": 0.00183243531516329, + "loss": 4.233404159545898, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.010381227359175682, + "learning_rate": 0.0018311330640501204, + "loss": 4.182013034820557, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.00900198519229889, + "learning_rate": 0.0018298305582342513, + "loss": 4.215488433837891, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.010198601521551609, + "learning_rate": 0.0018285277987650876, + "loss": 4.19389533996582, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.009965402074158192, + "learning_rate": 0.0018272247866922401, + "loss": 4.218005657196045, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.009299935773015022, + "learning_rate": 0.0018259215230655202, + "loss": 4.223684310913086, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.008977951481938362, + "learning_rate": 0.0018246180089349462, + "loss": 4.215778350830078, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.009261307306587696, + "learning_rate": 0.0018233142453507344, + "loss": 4.215066909790039, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.008322088047862053, + "learning_rate": 0.0018220102333633044, + "loss": 4.2188005447387695, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.008716144599020481, + "learning_rate": 0.0018207059740232739, + "loss": 4.187867641448975, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.008897020481526852, + "learning_rate": 0.0018194014683814625, + "loss": 4.218034267425537, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.007298174314200878, + "learning_rate": 0.0018180967174888857, + "loss": 4.2017011642456055, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.008491789922118187, + "learning_rate": 0.0018167917223967582, + "loss": 4.239873886108398, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.008413378149271011, + "learning_rate": 0.0018154864841564908, + "loss": 4.234771251678467, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.008141471073031425, + "learning_rate": 0.0018141810038196909, + "loss": 4.185944080352783, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.009480944834649563, + "learning_rate": 0.0018128752824381603, + "loss": 4.206586837768555, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.0106047997251153, + "learning_rate": 0.001811569321063894, + "loss": 4.235287666320801, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.010663514956831932, + "learning_rate": 0.0018102631207490838, + "loss": 4.204899311065674, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.011473523452877998, + "learning_rate": 0.001808956682546109, + "loss": 4.253668308258057, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01165992021560669, + "learning_rate": 0.0018076500075075456, + "loss": 4.209845542907715, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.011692410334944725, + "learning_rate": 0.0018063430966861564, + "loss": 4.250368595123291, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01061506848782301, + "learning_rate": 0.0018050359511348965, + "loss": 4.193821907043457, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.010510941036045551, + "learning_rate": 0.0018037285719069097, + "loss": 4.254327774047852, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.011587041430175304, + "learning_rate": 0.001802420960055528, + "loss": 4.2838945388793945, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.01257420051842928, + "learning_rate": 0.0018011131166342693, + "loss": 4.258241653442383, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.00936233438551426, + "learning_rate": 0.0017998050426968407, + "loss": 4.201854705810547, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.010093127377331257, + "learning_rate": 0.0017984967392971338, + "loss": 4.193051338195801, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.010032009333372116, + "learning_rate": 0.0017971882074892236, + "loss": 4.230212211608887, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.009455587714910507, + "learning_rate": 0.0017958794483273726, + "loss": 4.2157673835754395, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.009549238719046116, + "learning_rate": 0.001794570462866023, + "loss": 4.260272026062012, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.011059699580073357, + "learning_rate": 0.0017932612521598013, + "loss": 4.214818000793457, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.011653627268970013, + "learning_rate": 0.0017919518172635146, + "loss": 4.205112457275391, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.012438314035534859, + "learning_rate": 0.001790642159232151, + "loss": 4.179045677185059, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.011234673671424389, + "learning_rate": 0.0017893322791208784, + "loss": 4.239317417144775, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.010923845693469048, + "learning_rate": 0.0017880221779850434, + "loss": 4.230571269989014, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.009219309315085411, + "learning_rate": 0.001786711856880171, + "loss": 4.229068756103516, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.009849398396909237, + "learning_rate": 0.0017854013168619638, + "loss": 4.254772186279297, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.01117763202637434, + "learning_rate": 0.001784090558986299, + "loss": 4.252647399902344, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.010664612054824829, + "learning_rate": 0.0017827795843092316, + "loss": 4.204596519470215, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.00847950391471386, + "learning_rate": 0.0017814683938869895, + "loss": 4.211061954498291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.007305726408958435, + "learning_rate": 0.0017801569887759753, + "loss": 4.181221008300781, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.007970666512846947, + "learning_rate": 0.0017788453700327653, + "loss": 4.237855434417725, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.008873614482581615, + "learning_rate": 0.0017775335387141054, + "loss": 4.229188919067383, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.009691286832094193, + "learning_rate": 0.0017762214958769156, + "loss": 4.205491542816162, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.009360672906041145, + "learning_rate": 0.0017749092425782852, + "loss": 4.233466625213623, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.009895884431898594, + "learning_rate": 0.0017735967798754723, + "loss": 4.187763214111328, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.535672187805176, + "eval_runtime": 53.1736, + "eval_samples_per_second": 45.925, + "eval_steps_per_second": 1.448, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.00966129545122385, + "learning_rate": 0.0017722841088259045, + "loss": 4.1451735496521, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.010733380913734436, + "learning_rate": 0.0017709712304871774, + "loss": 4.1616363525390625, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.0117328567430377, + "learning_rate": 0.0017696581459170535, + "loss": 4.126331806182861, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.012640298344194889, + "learning_rate": 0.0017683448561734607, + "loss": 4.163730621337891, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.013074076734483242, + "learning_rate": 0.0017670313623144934, + "loss": 4.199502944946289, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.01132719125598669, + "learning_rate": 0.0017657176653984097, + "loss": 4.143380165100098, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.010008333250880241, + "learning_rate": 0.0017644037664836319, + "loss": 4.170320510864258, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.011090578511357307, + "learning_rate": 0.0017630896666287438, + "loss": 4.1571478843688965, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.011047456413507462, + "learning_rate": 0.0017617753668924924, + "loss": 4.101892948150635, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.010098653845489025, + "learning_rate": 0.0017604608683337857, + "loss": 4.154077529907227, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.010499860160052776, + "learning_rate": 0.00175914617201169, + "loss": 4.116868019104004, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.010939318686723709, + "learning_rate": 0.0017578312789854338, + "loss": 4.159144878387451, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.010622999630868435, + "learning_rate": 0.0017565161903144023, + "loss": 4.171411991119385, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.0108700692653656, + "learning_rate": 0.0017552009070581386, + "loss": 4.100748062133789, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.009794988669455051, + "learning_rate": 0.0017538854302763423, + "loss": 4.185161590576172, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.007750354241579771, + "learning_rate": 0.0017525697610288702, + "loss": 4.149083137512207, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.009923134930431843, + "learning_rate": 0.001751253900375732, + "loss": 4.205166816711426, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.010811141692101955, + "learning_rate": 0.0017499378493770936, + "loss": 4.131865978240967, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.010868688113987446, + "learning_rate": 0.0017486216090932733, + "loss": 4.148791790008545, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.011224126443266869, + "learning_rate": 0.001747305180584743, + "loss": 4.152503490447998, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.013262554071843624, + "learning_rate": 0.001745988564912124, + "loss": 4.164926528930664, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.013089599087834358, + "learning_rate": 0.001744671763136191, + "loss": 4.1626458168029785, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.01327646616846323, + "learning_rate": 0.0017433547763178665, + "loss": 4.189356803894043, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.014539042487740517, + "learning_rate": 0.0017420376055182233, + "loss": 4.176387786865234, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.013269098475575447, + "learning_rate": 0.001740720251798482, + "loss": 4.180716514587402, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.010922391898930073, + "learning_rate": 0.0017394027162200105, + "loss": 4.163928031921387, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.013355271890759468, + "learning_rate": 0.0017380849998443243, + "loss": 4.150012016296387, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.014149201102554798, + "learning_rate": 0.0017367671037330828, + "loss": 4.173713207244873, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01354879979044199, + "learning_rate": 0.0017354490289480918, + "loss": 4.150563716888428, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.013788704760372639, + "learning_rate": 0.0017341307765512998, + "loss": 4.213858604431152, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.011597222648561, + "learning_rate": 0.0017328123476047994, + "loss": 4.166989326477051, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.01055675558745861, + "learning_rate": 0.001731493743170824, + "loss": 4.172226428985596, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.011600397527217865, + "learning_rate": 0.0017301749643117506, + "loss": 4.1656951904296875, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.01114651933312416, + "learning_rate": 0.0017288560120900943, + "loss": 4.189825057983398, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.009811583906412125, + "learning_rate": 0.0017275368875685126, + "loss": 4.181164741516113, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.009758968837559223, + "learning_rate": 0.0017262175918097992, + "loss": 4.163188934326172, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.00999707542359829, + "learning_rate": 0.001724898125876887, + "loss": 4.184117317199707, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.009528180584311485, + "learning_rate": 0.001723578490832846, + "loss": 4.16407585144043, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.009717358276247978, + "learning_rate": 0.0017222586877408826, + "loss": 4.209892272949219, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.010909050703048706, + "learning_rate": 0.0017209387176643377, + "loss": 4.19977331161499, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.011234750039875507, + "learning_rate": 0.0017196185816666875, + "loss": 4.19363260269165, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.010300721041858196, + "learning_rate": 0.0017182982808115425, + "loss": 4.142427921295166, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.00941646471619606, + "learning_rate": 0.001716977816162644, + "loss": 4.159372329711914, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.009304054081439972, + "learning_rate": 0.0017156571887838679, + "loss": 4.207922458648682, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.009496180340647697, + "learning_rate": 0.0017143363997392184, + "loss": 4.137894630432129, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.009257436729967594, + "learning_rate": 0.0017130154500928326, + "loss": 4.198753833770752, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.00900586973875761, + "learning_rate": 0.0017116943409089746, + "loss": 4.160684585571289, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.009398494847118855, + "learning_rate": 0.00171037307325204, + "loss": 4.1819353103637695, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.008633013814687729, + "learning_rate": 0.0017090516481865483, + "loss": 4.176833152770996, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.00898425467312336, + "learning_rate": 0.0017077300667771484, + "loss": 4.190657138824463, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.008182398043572903, + "learning_rate": 0.0017064083300886155, + "loss": 4.173429489135742, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.009589143097400665, + "learning_rate": 0.0017050864391858478, + "loss": 4.1559014320373535, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.010304383002221584, + "learning_rate": 0.0017037643951338698, + "loss": 4.155453681945801, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.009943748824298382, + "learning_rate": 0.0017024421989978275, + "loss": 4.119734287261963, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.009238033555448055, + "learning_rate": 0.001701119851842992, + "loss": 4.188974857330322, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.008058690465986729, + "learning_rate": 0.0016997973547347531, + "loss": 4.19934606552124, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.008485953323543072, + "learning_rate": 0.0016984747087386233, + "loss": 4.165615558624268, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.008476605638861656, + "learning_rate": 0.0016971519149202347, + "loss": 4.21021842956543, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.008588708005845547, + "learning_rate": 0.0016958289743453385, + "loss": 4.208064556121826, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.009223178029060364, + "learning_rate": 0.0016945058880798037, + "loss": 4.158413887023926, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.009196167811751366, + "learning_rate": 0.0016931826571896166, + "loss": 4.203187465667725, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.009488921612501144, + "learning_rate": 0.001691859282740882, + "loss": 4.178227424621582, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.009253161028027534, + "learning_rate": 0.0016905357657998167, + "loss": 4.131230354309082, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.010069048963487148, + "learning_rate": 0.0016892121074327559, + "loss": 4.210734844207764, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.009945095516741276, + "learning_rate": 0.001687888308706147, + "loss": 4.184042930603027, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.010483640246093273, + "learning_rate": 0.0016865643706865505, + "loss": 4.156919479370117, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.011560715734958649, + "learning_rate": 0.0016852402944406391, + "loss": 4.185520172119141, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.010001277551054955, + "learning_rate": 0.0016839160810351978, + "loss": 4.226828575134277, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.010335934348404408, + "learning_rate": 0.0016825917315371202, + "loss": 4.183531284332275, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.011322776786983013, + "learning_rate": 0.001681267247013412, + "loss": 4.179635047912598, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.009663194417953491, + "learning_rate": 0.001679942628531186, + "loss": 4.165654182434082, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.010590246878564358, + "learning_rate": 0.0016786178771576643, + "loss": 4.1539459228515625, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.511386394500732, + "eval_runtime": 51.7751, + "eval_samples_per_second": 47.166, + "eval_steps_per_second": 1.487, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.01254354789853096, + "learning_rate": 0.0016772929939601728, + "loss": 4.104002475738525, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.01300987508147955, + "learning_rate": 0.0016759679800061483, + "loss": 4.127395153045654, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.013325816951692104, + "learning_rate": 0.0016746428363631299, + "loss": 4.131229400634766, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.014302870258688927, + "learning_rate": 0.001673317564098761, + "loss": 4.079358100891113, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.013776068575680256, + "learning_rate": 0.0016719921642807905, + "loss": 4.132065773010254, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.010788560844957829, + "learning_rate": 0.0016706666379770689, + "loss": 4.134626388549805, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.009766630828380585, + "learning_rate": 0.0016693409862555493, + "loss": 4.1645426750183105, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.011760029010474682, + "learning_rate": 0.0016680152101842842, + "loss": 4.110664367675781, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.011047836393117905, + "learning_rate": 0.0016666893108314288, + "loss": 4.129222869873047, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.009989139623939991, + "learning_rate": 0.0016653632892652344, + "loss": 4.076835632324219, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.010033783502876759, + "learning_rate": 0.001664037146554054, + "loss": 4.11555290222168, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.00996427796781063, + "learning_rate": 0.001662710883766337, + "loss": 4.131463527679443, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.01044023409485817, + "learning_rate": 0.0016613845019706288, + "loss": 4.098750114440918, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.010033872909843922, + "learning_rate": 0.0016600580022355712, + "loss": 4.161447048187256, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.009329567663371563, + "learning_rate": 0.0016587313856299012, + "loss": 4.146491527557373, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.010061589069664478, + "learning_rate": 0.0016574046532224504, + "loss": 4.107517719268799, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.009491910226643085, + "learning_rate": 0.0016560778060821424, + "loss": 4.098226547241211, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.008529345504939556, + "learning_rate": 0.0016547508452779943, + "loss": 4.1247711181640625, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.00936342403292656, + "learning_rate": 0.0016534237718791136, + "loss": 4.120573043823242, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.010255641303956509, + "learning_rate": 0.0016520965869547008, + "loss": 4.089067459106445, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.011638758704066277, + "learning_rate": 0.001650769291574044, + "loss": 4.074847221374512, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.014707857742905617, + "learning_rate": 0.0016494418868065214, + "loss": 4.153377532958984, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.01650111749768257, + "learning_rate": 0.0016481143737215986, + "loss": 4.134062767028809, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.014833372086286545, + "learning_rate": 0.00164678675338883, + "loss": 4.1344828605651855, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.012257886119186878, + "learning_rate": 0.0016454590268778538, + "loss": 4.098813533782959, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.013508322648704052, + "learning_rate": 0.0016441311952583959, + "loss": 4.106297492980957, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.01256900280714035, + "learning_rate": 0.001642803259600267, + "loss": 4.143631935119629, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.009885406121611595, + "learning_rate": 0.00164147522097336, + "loss": 4.120885848999023, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.012004674412310123, + "learning_rate": 0.0016401470804476518, + "loss": 4.089243412017822, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01224028691649437, + "learning_rate": 0.0016388188390932012, + "loss": 4.077765464782715, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.012511290609836578, + "learning_rate": 0.001637490497980149, + "loss": 4.161346435546875, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.010701716877520084, + "learning_rate": 0.0016361620581787143, + "loss": 4.10698938369751, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.010713113471865654, + "learning_rate": 0.001634833520759198, + "loss": 4.114283561706543, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.011018907651305199, + "learning_rate": 0.001633504886791978, + "loss": 4.134284019470215, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.011279584839940071, + "learning_rate": 0.0016321761573475117, + "loss": 4.1456804275512695, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.010642931796610355, + "learning_rate": 0.001630847333496332, + "loss": 4.160918235778809, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.011069678701460361, + "learning_rate": 0.0016295184163090482, + "loss": 4.136838912963867, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.011007166467607021, + "learning_rate": 0.0016281894068563443, + "loss": 4.129153728485107, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.010055703110992908, + "learning_rate": 0.0016268603062089802, + "loss": 4.138665676116943, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.009839230217039585, + "learning_rate": 0.0016255311154377875, + "loss": 4.1366496086120605, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.011191432364284992, + "learning_rate": 0.0016242018356136714, + "loss": 4.129880905151367, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.012130341492593288, + "learning_rate": 0.0016228724678076095, + "loss": 4.1425981521606445, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.012707795947790146, + "learning_rate": 0.0016215430130906482, + "loss": 4.115452766418457, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.012707809917628765, + "learning_rate": 0.001620213472533906, + "loss": 4.122758865356445, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.01182397548109293, + "learning_rate": 0.001618883847208569, + "loss": 4.140275955200195, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.010865340940654278, + "learning_rate": 0.001617554138185893, + "loss": 4.137759208679199, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.010973592288792133, + "learning_rate": 0.0016162243465371998, + "loss": 4.138947010040283, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.011247226968407631, + "learning_rate": 0.001614894473333879, + "loss": 4.115516662597656, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.010837487876415253, + "learning_rate": 0.0016135645196473853, + "loss": 4.160427093505859, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.009096396155655384, + "learning_rate": 0.0016122344865492385, + "loss": 4.1200032234191895, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.010671247728168964, + "learning_rate": 0.0016109043751110223, + "loss": 4.097919464111328, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.0107562355697155, + "learning_rate": 0.0016095741864043831, + "loss": 4.096290588378906, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.008909327909350395, + "learning_rate": 0.0016082439215010305, + "loss": 4.171017646789551, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.010305124334990978, + "learning_rate": 0.0016069135814727343, + "loss": 4.1450276374816895, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.011235613375902176, + "learning_rate": 0.0016055831673913255, + "loss": 4.134704113006592, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.00825632456690073, + "learning_rate": 0.0016042526803286949, + "loss": 4.107443332672119, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.009056378155946732, + "learning_rate": 0.0016029221213567921, + "loss": 4.132656097412109, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.008835164830088615, + "learning_rate": 0.001601591491547624, + "loss": 4.124729633331299, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.00915333442389965, + "learning_rate": 0.0016002607919732557, + "loss": 4.119500160217285, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.008532539941370487, + "learning_rate": 0.0015989300237058066, + "loss": 4.141661643981934, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.009952311404049397, + "learning_rate": 0.0015975991878174533, + "loss": 4.140929222106934, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.008920297026634216, + "learning_rate": 0.0015962682853804268, + "loss": 4.155877113342285, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.008770049549639225, + "learning_rate": 0.0015949373174670104, + "loss": 4.126394271850586, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.0078084105625748634, + "learning_rate": 0.0015936062851495411, + "loss": 4.128876209259033, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.00816641841083765, + "learning_rate": 0.0015922751895004078, + "loss": 4.137410640716553, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.008895459584891796, + "learning_rate": 0.0015909440315920505, + "loss": 4.07957649230957, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.008551528677344322, + "learning_rate": 0.001589612812496958, + "loss": 4.104213714599609, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.008325488306581974, + "learning_rate": 0.0015882815332876707, + "loss": 4.1584792137146, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.009341313503682613, + "learning_rate": 0.0015869501950367744, + "loss": 4.1417436599731445, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.00999050959944725, + "learning_rate": 0.0015856187988169067, + "loss": 4.136049270629883, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.008889932185411453, + "learning_rate": 0.0015842873457007484, + "loss": 4.16257381439209, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.009109573438763618, + "learning_rate": 0.0015829558367610268, + "loss": 4.171441555023193, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.485942840576172, + "eval_runtime": 52.1051, + "eval_samples_per_second": 46.867, + "eval_steps_per_second": 1.478, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.009881424717605114, + "learning_rate": 0.001581624273070515, + "loss": 4.059608459472656, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.011156721040606499, + "learning_rate": 0.0015802926557020294, + "loss": 4.047014236450195, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.012326992116868496, + "learning_rate": 0.0015789609857284309, + "loss": 4.06541109085083, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.01332208700478077, + "learning_rate": 0.0015776292642226213, + "loss": 4.042714595794678, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.012720128521323204, + "learning_rate": 0.0015762974922575446, + "loss": 4.098389625549316, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.010723215527832508, + "learning_rate": 0.0015749656709061854, + "loss": 4.059185981750488, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.010605281218886375, + "learning_rate": 0.001573633801241569, + "loss": 4.099538326263428, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.01101530622690916, + "learning_rate": 0.0015723018843367576, + "loss": 4.095674514770508, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.01135838869959116, + "learning_rate": 0.0015709699212648532, + "loss": 4.070535659790039, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.012119271792471409, + "learning_rate": 0.0015696379130989936, + "loss": 4.060605049133301, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.012783961370587349, + "learning_rate": 0.001568305860912355, + "loss": 4.075931072235107, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.011486220173537731, + "learning_rate": 0.0015669737657781461, + "loss": 4.075855255126953, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.01037552859634161, + "learning_rate": 0.0015656416287696136, + "loss": 4.087729454040527, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.009802743792533875, + "learning_rate": 0.0015643094509600354, + "loss": 4.0855607986450195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.010804438032209873, + "learning_rate": 0.001562977233422723, + "loss": 4.064528465270996, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.009505728259682655, + "learning_rate": 0.0015616449772310206, + "loss": 4.036775588989258, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.010634485632181168, + "learning_rate": 0.0015603126834583024, + "loss": 4.069923400878906, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.011392096057534218, + "learning_rate": 0.0015589803531779736, + "loss": 4.0812177658081055, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.011851722374558449, + "learning_rate": 0.0015576479874634683, + "loss": 4.09183406829834, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.011205444112420082, + "learning_rate": 0.0015563155873882502, + "loss": 4.07825231552124, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.009676349349319935, + "learning_rate": 0.0015549831540258093, + "loss": 4.052826404571533, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.009376165457069874, + "learning_rate": 0.0015536506884496638, + "loss": 4.02100133895874, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.00936955213546753, + "learning_rate": 0.0015523181917333564, + "loss": 4.049526214599609, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.00937784556299448, + "learning_rate": 0.0015509856649504564, + "loss": 4.096356391906738, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.010014008730649948, + "learning_rate": 0.0015496531091745555, + "loss": 4.03254508972168, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.009909634478390217, + "learning_rate": 0.0015483205254792705, + "loss": 4.067630767822266, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.009319501928985119, + "learning_rate": 0.0015469879149382399, + "loss": 4.094167709350586, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.008634721860289574, + "learning_rate": 0.0015456552786251244, + "loss": 4.122404098510742, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.009412454441189766, + "learning_rate": 0.0015443226176136049, + "loss": 4.081546783447266, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.010102108120918274, + "learning_rate": 0.0015429899329773814, + "loss": 4.049839019775391, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.010921654291450977, + "learning_rate": 0.0015416572257901746, + "loss": 4.097097396850586, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.010607713833451271, + "learning_rate": 0.0015403244971257226, + "loss": 4.084481239318848, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.011887960135936737, + "learning_rate": 0.0015389917480577806, + "loss": 4.074968338012695, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01089888159185648, + "learning_rate": 0.0015376589796601203, + "loss": 4.062132835388184, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.009934183210134506, + "learning_rate": 0.0015363261930065298, + "loss": 4.128092288970947, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.00878534372895956, + "learning_rate": 0.0015349933891708105, + "loss": 4.069051742553711, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.009928155690431595, + "learning_rate": 0.0015336605692267788, + "loss": 4.073912620544434, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.011091411113739014, + "learning_rate": 0.0015323277342482637, + "loss": 4.125944137573242, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.010314172133803368, + "learning_rate": 0.0015309948853091055, + "loss": 4.062620162963867, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.010423967614769936, + "learning_rate": 0.001529662023483158, + "loss": 4.101737976074219, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.009822997264564037, + "learning_rate": 0.0015283291498442817, + "loss": 4.064520359039307, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.009757095016539097, + "learning_rate": 0.001526996265466351, + "loss": 4.100624084472656, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.008650688454508781, + "learning_rate": 0.001525663371423246, + "loss": 4.109685897827148, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.007378234062343836, + "learning_rate": 0.0015243304687888553, + "loss": 4.044726848602295, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.008194336667656898, + "learning_rate": 0.001522997558637074, + "loss": 4.097923278808594, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.009764892049133778, + "learning_rate": 0.0015216646420418048, + "loss": 4.102148056030273, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.009242868050932884, + "learning_rate": 0.001520331720076953, + "loss": 4.068563461303711, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.010409324429929256, + "learning_rate": 0.001518998793816431, + "loss": 4.065944194793701, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.009215258061885834, + "learning_rate": 0.0015176658643341527, + "loss": 4.10223388671875, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.008065393194556236, + "learning_rate": 0.0015163329327040366, + "loss": 4.099432945251465, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.008498288691043854, + "learning_rate": 0.001515, + "loss": 4.109854698181152, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.008919150568544865, + "learning_rate": 0.0015136670672959633, + "loss": 4.128247261047363, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.009104429744184017, + "learning_rate": 0.0015123341356658474, + "loss": 4.068622589111328, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.008558365516364574, + "learning_rate": 0.0015110012061835692, + "loss": 4.114214897155762, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.007865076884627342, + "learning_rate": 0.0015096682799230477, + "loss": 4.11155366897583, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.007848534733057022, + "learning_rate": 0.0015083353579581958, + "loss": 4.045694351196289, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.008238030597567558, + "learning_rate": 0.0015070024413629264, + "loss": 4.093930721282959, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.007536398712545633, + "learning_rate": 0.0015056695312111448, + "loss": 4.106235504150391, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.008156892843544483, + "learning_rate": 0.001504336628576754, + "loss": 4.13060998916626, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.008144439198076725, + "learning_rate": 0.0015030037345336487, + "loss": 4.091304779052734, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.007757358253002167, + "learning_rate": 0.001501670850155718, + "loss": 4.069276809692383, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.00860623549669981, + "learning_rate": 0.0015003379765168425, + "loss": 4.107658386230469, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.008614457212388515, + "learning_rate": 0.0014990051146908949, + "loss": 4.106705665588379, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.009120592847466469, + "learning_rate": 0.001497672265751737, + "loss": 4.0891923904418945, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.009446980431675911, + "learning_rate": 0.0014963394307732213, + "loss": 4.098847389221191, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.010494766756892204, + "learning_rate": 0.0014950066108291894, + "loss": 4.115325450897217, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.010057658888399601, + "learning_rate": 0.0014936738069934706, + "loss": 4.099984169006348, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.009071771055459976, + "learning_rate": 0.0014923410203398796, + "loss": 4.091543674468994, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.009937075898051262, + "learning_rate": 0.00149100825194222, + "loss": 4.086492538452148, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.010222986340522766, + "learning_rate": 0.0014896755028742778, + "loss": 4.063841819763184, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.010513762012124062, + "learning_rate": 0.001488342774209826, + "loss": 4.11991548538208, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.009530222974717617, + "learning_rate": 0.0014870100670226192, + "loss": 4.043334007263184, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.467459678649902, + "eval_runtime": 54.1628, + "eval_samples_per_second": 45.086, + "eval_steps_per_second": 1.422, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.010274187661707401, + "learning_rate": 0.0014856773823863955, + "loss": 4.049561500549316, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.011448442004621029, + "learning_rate": 0.001484344721374876, + "loss": 4.019484996795654, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.011361192911863327, + "learning_rate": 0.0014830120850617599, + "loss": 4.0429887771606445, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.010250156745314598, + "learning_rate": 0.0014816794745207299, + "loss": 4.012682914733887, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.010024400427937508, + "learning_rate": 0.0014803468908254449, + "loss": 3.979409694671631, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.01153535582125187, + "learning_rate": 0.0014790143350495446, + "loss": 4.045780181884766, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.012632817961275578, + "learning_rate": 0.001477681808266644, + "loss": 4.057641983032227, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.013758293353021145, + "learning_rate": 0.0014763493115503366, + "loss": 4.061607360839844, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.014041583053767681, + "learning_rate": 0.001475016845974191, + "loss": 4.035029411315918, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.013629456982016563, + "learning_rate": 0.0014736844126117497, + "loss": 4.062169075012207, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.011023213155567646, + "learning_rate": 0.0014723520125365323, + "loss": 4.02731990814209, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.01001750584691763, + "learning_rate": 0.0014710196468220266, + "loss": 4.019882678985596, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.008321646600961685, + "learning_rate": 0.001469687316541698, + "loss": 4.005446434020996, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.008935946971178055, + "learning_rate": 0.0014683550227689796, + "loss": 4.017102241516113, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.009137130342423916, + "learning_rate": 0.001467022766577277, + "loss": 4.024102210998535, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.00805022194981575, + "learning_rate": 0.001465690549039965, + "loss": 3.9935250282287598, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.009204058907926083, + "learning_rate": 0.0014643583712303865, + "loss": 4.042818546295166, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.009812815114855766, + "learning_rate": 0.001463026234221854, + "loss": 4.0252485275268555, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.009955782443284988, + "learning_rate": 0.0014616941390876454, + "loss": 3.9839768409729004, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.009374645538628101, + "learning_rate": 0.001460362086901007, + "loss": 4.040245056152344, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.010008739307522774, + "learning_rate": 0.0014590300787351471, + "loss": 4.0306596755981445, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.009014200419187546, + "learning_rate": 0.001457698115663243, + "loss": 4.028713226318359, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.008381523191928864, + "learning_rate": 0.0014563661987584313, + "loss": 4.024718284606934, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.008627948351204395, + "learning_rate": 0.0014550343290938143, + "loss": 4.02303409576416, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.010257705114781857, + "learning_rate": 0.0014537025077424556, + "loss": 4.061058044433594, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.010759209282696247, + "learning_rate": 0.0014523707357773788, + "loss": 4.064718246459961, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.009966139681637287, + "learning_rate": 0.0014510390142715697, + "loss": 3.9919447898864746, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.00997584406286478, + "learning_rate": 0.001449707344297971, + "loss": 4.0431599617004395, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.010846944525837898, + "learning_rate": 0.0014483757269294858, + "loss": 4.0222697257995605, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.00991352554410696, + "learning_rate": 0.0014470441632389735, + "loss": 4.059438705444336, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.008994397707283497, + "learning_rate": 0.0014457126542992518, + "loss": 4.021494388580322, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.009567267261445522, + "learning_rate": 0.0014443812011830934, + "loss": 4.070810317993164, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.009199011139571667, + "learning_rate": 0.0014430498049632255, + "loss": 4.020938873291016, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.010440350510179996, + "learning_rate": 0.0014417184667123303, + "loss": 4.062529563903809, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.009962482377886772, + "learning_rate": 0.0014403871875030424, + "loss": 4.0687150955200195, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.009389379993081093, + "learning_rate": 0.0014390559684079502, + "loss": 4.055185794830322, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.011469941586256027, + "learning_rate": 0.0014377248104995924, + "loss": 4.041023254394531, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.011034650728106499, + "learning_rate": 0.0014363937148504588, + "loss": 4.054254531860352, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.010651151649653912, + "learning_rate": 0.0014350626825329897, + "loss": 4.037753582000732, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.010747930034995079, + "learning_rate": 0.0014337317146195733, + "loss": 4.059938907623291, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.00987576600164175, + "learning_rate": 0.0014324008121825468, + "loss": 4.075366973876953, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.008504371158778667, + "learning_rate": 0.0014310699762941936, + "loss": 4.026755332946777, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.009445052593946457, + "learning_rate": 0.001429739208026745, + "loss": 4.034431457519531, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.010560023598372936, + "learning_rate": 0.0014284085084523763, + "loss": 4.101158618927002, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.009920580312609673, + "learning_rate": 0.0014270778786432078, + "loss": 4.041321754455566, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.009854268282651901, + "learning_rate": 0.0014257473196713053, + "loss": 4.049882411956787, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.009825135581195354, + "learning_rate": 0.0014244168326086745, + "loss": 4.067440986633301, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.010506784543395042, + "learning_rate": 0.0014230864185272663, + "loss": 4.042476654052734, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.009527767077088356, + "learning_rate": 0.0014217560784989696, + "loss": 4.055248260498047, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.01013826485723257, + "learning_rate": 0.001420425813595617, + "loss": 4.05246639251709, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.011020111851394176, + "learning_rate": 0.0014190956248889778, + "loss": 4.032987117767334, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.010647117160260677, + "learning_rate": 0.0014177655134507614, + "loss": 4.093589782714844, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.009484780952334404, + "learning_rate": 0.001416435480352615, + "loss": 4.017437934875488, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.0114199910312891, + "learning_rate": 0.0014151055266661211, + "loss": 4.080643653869629, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.011137664318084717, + "learning_rate": 0.0014137756534628007, + "loss": 4.069380760192871, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.008972682990133762, + "learning_rate": 0.0014124458618141073, + "loss": 4.051383018493652, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.010311643593013287, + "learning_rate": 0.0014111161527914317, + "loss": 4.048953533172607, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.01113453321158886, + "learning_rate": 0.0014097865274660944, + "loss": 4.063544273376465, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.01091703586280346, + "learning_rate": 0.0014084569869093518, + "loss": 4.006721496582031, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.010672449134290218, + "learning_rate": 0.0014071275321923906, + "loss": 4.020220756530762, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.009682527743279934, + "learning_rate": 0.001405798164386328, + "loss": 4.057307243347168, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.009808676317334175, + "learning_rate": 0.0014044688845622126, + "loss": 4.105412483215332, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.010351189412176609, + "learning_rate": 0.0014031396937910202, + "loss": 4.066454887390137, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.010038275271654129, + "learning_rate": 0.001401810593143656, + "loss": 4.056752681732178, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01016094908118248, + "learning_rate": 0.001400481583690952, + "loss": 4.097933292388916, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.008683547377586365, + "learning_rate": 0.0013991526665036679, + "loss": 4.0386152267456055, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.009556332603096962, + "learning_rate": 0.0013978238426524882, + "loss": 4.05835485458374, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.009597044438123703, + "learning_rate": 0.001396495113208022, + "loss": 4.052363872528076, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.010021139867603779, + "learning_rate": 0.0013951664792408026, + "loss": 4.083431243896484, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.010611358098685741, + "learning_rate": 0.0013938379418212863, + "loss": 4.042330265045166, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.009132112376391888, + "learning_rate": 0.001392509502019852, + "loss": 4.058825969696045, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.009808709844946861, + "learning_rate": 0.0013911811609067992, + "loss": 4.093010902404785, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.4473161697387695, + "eval_runtime": 51.7738, + "eval_samples_per_second": 47.167, + "eval_steps_per_second": 1.487, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.011859052814543247, + "learning_rate": 0.0013898529195523488, + "loss": 3.9572651386260986, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.013210238888859749, + "learning_rate": 0.0013885247790266406, + "loss": 3.9812350273132324, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.012981324456632137, + "learning_rate": 0.001387196740399733, + "loss": 3.987725019454956, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.01424382347613573, + "learning_rate": 0.001385868804741604, + "loss": 3.9719889163970947, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.0154642965644598, + "learning_rate": 0.0013845409731221465, + "loss": 3.984248161315918, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.014779271557927132, + "learning_rate": 0.0013832132466111706, + "loss": 4.002912521362305, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.012921188026666641, + "learning_rate": 0.0013818856262784013, + "loss": 3.9588499069213867, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01165325753390789, + "learning_rate": 0.0013805581131934792, + "loss": 3.993288993835449, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.008908954448997974, + "learning_rate": 0.0013792307084259563, + "loss": 3.987151861190796, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.012040507979691029, + "learning_rate": 0.0013779034130452991, + "loss": 3.9675703048706055, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.013464349322021008, + "learning_rate": 0.0013765762281208868, + "loss": 3.9901437759399414, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.009626250714063644, + "learning_rate": 0.0013752491547220063, + "loss": 3.9373281002044678, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.00897257961332798, + "learning_rate": 0.0013739221939178582, + "loss": 4.00804328918457, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.010689414106309414, + "learning_rate": 0.00137259534677755, + "loss": 4.020804405212402, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.010462156496942043, + "learning_rate": 0.0013712686143700988, + "loss": 4.028973579406738, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.009135857224464417, + "learning_rate": 0.001369941997764429, + "loss": 3.9798216819763184, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.009298899210989475, + "learning_rate": 0.0013686154980293713, + "loss": 4.018845081329346, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.00979928020387888, + "learning_rate": 0.0013672891162336636, + "loss": 3.9946775436401367, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.008175542578101158, + "learning_rate": 0.0013659628534459457, + "loss": 3.981858491897583, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.008302642032504082, + "learning_rate": 0.0013646367107347662, + "loss": 3.991128444671631, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.008516009896993637, + "learning_rate": 0.0013633106891685718, + "loss": 4.004990100860596, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.007936482317745686, + "learning_rate": 0.0013619847898157164, + "loss": 3.995734214782715, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.008317600004374981, + "learning_rate": 0.001360659013744451, + "loss": 4.040653228759766, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.008481282740831375, + "learning_rate": 0.001359333362022931, + "loss": 3.969860553741455, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.008944948203861713, + "learning_rate": 0.0013580078357192096, + "loss": 3.9908909797668457, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.008897850289940834, + "learning_rate": 0.001356682435901239, + "loss": 4.04786491394043, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.009527456946671009, + "learning_rate": 0.001355357163636871, + "loss": 4.017423629760742, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.011714582331478596, + "learning_rate": 0.001354032019993852, + "loss": 3.9990692138671875, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.011083807796239853, + "learning_rate": 0.0013527070060398274, + "loss": 3.995265483856201, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.009878288954496384, + "learning_rate": 0.0013513821228423363, + "loss": 4.0364532470703125, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.009727329947054386, + "learning_rate": 0.0013500573714688136, + "loss": 4.035953521728516, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.009482033550739288, + "learning_rate": 0.0013487327529865879, + "loss": 4.015855312347412, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.009037062525749207, + "learning_rate": 0.0013474082684628798, + "loss": 4.024164199829102, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.009094092063605785, + "learning_rate": 0.0013460839189648028, + "loss": 4.037765979766846, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.008230654522776604, + "learning_rate": 0.0013447597055593612, + "loss": 4.006317615509033, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.008491289801895618, + "learning_rate": 0.0013434356293134501, + "loss": 4.042603969573975, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.008907693438231945, + "learning_rate": 0.0013421116912938533, + "loss": 4.008098602294922, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.008897818624973297, + "learning_rate": 0.0013407878925672438, + "loss": 3.9892945289611816, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.00957555789500475, + "learning_rate": 0.0013394642342001832, + "loss": 4.046009063720703, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.009948568418622017, + "learning_rate": 0.0013381407172591185, + "loss": 3.9760642051696777, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.01014955248683691, + "learning_rate": 0.0013368173428103835, + "loss": 4.023215293884277, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.008578895591199398, + "learning_rate": 0.0013354941119201966, + "loss": 4.002819061279297, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.008729875087738037, + "learning_rate": 0.001334171025654662, + "loss": 4.0340657234191895, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.009556280449032784, + "learning_rate": 0.0013328480850797655, + "loss": 3.983076333999634, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.009733820334076881, + "learning_rate": 0.0013315252912613766, + "loss": 4.066578388214111, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.009025252424180508, + "learning_rate": 0.0013302026452652474, + "loss": 3.9980697631835938, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.008485320024192333, + "learning_rate": 0.001328880148157008, + "loss": 4.018826961517334, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.009539696387946606, + "learning_rate": 0.0013275578010021726, + "loss": 4.011737823486328, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.010881016030907631, + "learning_rate": 0.0013262356048661305, + "loss": 4.007291793823242, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.010362236760556698, + "learning_rate": 0.0013249135608141523, + "loss": 3.9929394721984863, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.010296608321368694, + "learning_rate": 0.0013235916699113849, + "loss": 4.0015130043029785, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.011211174540221691, + "learning_rate": 0.0013222699332228517, + "loss": 4.005324363708496, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.009743994101881981, + "learning_rate": 0.0013209483518134525, + "loss": 4.0278143882751465, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.008563914336264133, + "learning_rate": 0.0013196269267479607, + "loss": 3.9997801780700684, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.009078902192413807, + "learning_rate": 0.0013183056590910255, + "loss": 4.026200294494629, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.009618895128369331, + "learning_rate": 0.0013169845499071675, + "loss": 4.027050018310547, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.008920644409954548, + "learning_rate": 0.0013156636002607822, + "loss": 4.000490665435791, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.008849202655255795, + "learning_rate": 0.0013143428112161323, + "loss": 4.044985294342041, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.009303621016442776, + "learning_rate": 0.001313022183837356, + "loss": 4.046281814575195, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.01076227892190218, + "learning_rate": 0.0013117017191884579, + "loss": 4.030576705932617, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.010400841943919659, + "learning_rate": 0.0013103814183333127, + "loss": 3.980081558227539, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.009459166787564754, + "learning_rate": 0.001309061282335663, + "loss": 4.014824867248535, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.009054969064891338, + "learning_rate": 0.001307741312259118, + "loss": 4.0485944747924805, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.009003991261124611, + "learning_rate": 0.0013064215091671545, + "loss": 3.9990320205688477, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.008959448896348476, + "learning_rate": 0.001305101874123113, + "loss": 4.038620948791504, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.009102243930101395, + "learning_rate": 0.0013037824081902015, + "loss": 4.030906677246094, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.008599337190389633, + "learning_rate": 0.0013024631124314877, + "loss": 4.02648401260376, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.008771977387368679, + "learning_rate": 0.0013011439879099056, + "loss": 4.027151107788086, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.009082595817744732, + "learning_rate": 0.00129982503568825, + "loss": 4.0194478034973145, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.009025354869663715, + "learning_rate": 0.0012985062568291763, + "loss": 4.044382095336914, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.00875102635473013, + "learning_rate": 0.0012971876523952014, + "loss": 4.0257568359375, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.009312704205513, + "learning_rate": 0.0012958692234487006, + "loss": 3.9964494705200195, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.431707859039307, + "eval_runtime": 50.7311, + "eval_samples_per_second": 48.136, + "eval_steps_per_second": 1.518, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.00992550328373909, + "learning_rate": 0.0012945509710519088, + "loss": 3.9253792762756348, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.009920453652739525, + "learning_rate": 0.0012932328962669173, + "loss": 3.9438819885253906, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.009437218308448792, + "learning_rate": 0.0012919150001556756, + "loss": 3.960963249206543, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.009616893716156483, + "learning_rate": 0.0012905972837799896, + "loss": 3.969301223754883, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.009107022546231747, + "learning_rate": 0.0012892797482015185, + "loss": 3.9626078605651855, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.011317167431116104, + "learning_rate": 0.0012879623944817772, + "loss": 3.946483850479126, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.010536770336329937, + "learning_rate": 0.001286645223682134, + "loss": 3.959566116333008, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.009349033236503601, + "learning_rate": 0.0012853282368638097, + "loss": 3.964435338973999, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.009950080886483192, + "learning_rate": 0.0012840114350878762, + "loss": 3.9729413986206055, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.010833023115992546, + "learning_rate": 0.001282694819415257, + "loss": 4.006421089172363, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.011538525111973286, + "learning_rate": 0.0012813783909067266, + "loss": 3.9558355808258057, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.010671315714716911, + "learning_rate": 0.0012800621506229065, + "loss": 4.0052337646484375, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.010835555382072926, + "learning_rate": 0.0012787460996242686, + "loss": 3.958641767501831, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.01082367543131113, + "learning_rate": 0.0012774302389711304, + "loss": 3.9818902015686035, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.009194346144795418, + "learning_rate": 0.0012761145697236583, + "loss": 3.9688854217529297, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.007859935984015465, + "learning_rate": 0.001274799092941862, + "loss": 3.9730639457702637, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.008534743450582027, + "learning_rate": 0.0012734838096855978, + "loss": 3.975003719329834, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.009369445964694023, + "learning_rate": 0.0012721687210145666, + "loss": 3.972907304763794, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.0090147340670228, + "learning_rate": 0.0012708538279883098, + "loss": 3.964256525039673, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.007745139300823212, + "learning_rate": 0.0012695391316662151, + "loss": 3.9504833221435547, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.008920101448893547, + "learning_rate": 0.0012682246331075075, + "loss": 3.9735097885131836, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.008987226523458958, + "learning_rate": 0.0012669103333712568, + "loss": 3.9482927322387695, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.00841289572417736, + "learning_rate": 0.0012655962335163685, + "loss": 3.965221881866455, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.008016116917133331, + "learning_rate": 0.0012642823346015902, + "loss": 3.9390206336975098, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.008470152504742146, + "learning_rate": 0.0012629686376855068, + "loss": 3.9257445335388184, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.008637234568595886, + "learning_rate": 0.0012616551438265394, + "loss": 3.9999356269836426, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.008036923594772816, + "learning_rate": 0.0012603418540829473, + "loss": 3.9755172729492188, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.008908843621611595, + "learning_rate": 0.0012590287695128232, + "loss": 3.9248545169830322, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.008743193000555038, + "learning_rate": 0.001257715891174096, + "loss": 3.9619390964508057, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.00834199320524931, + "learning_rate": 0.001256403220124528, + "loss": 3.961569309234619, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.009148893877863884, + "learning_rate": 0.001255090757421715, + "loss": 3.972104549407959, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.00925084762275219, + "learning_rate": 0.0012537785041230843, + "loss": 4.000775337219238, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.008659154176712036, + "learning_rate": 0.0012524664612858947, + "loss": 3.954564332962036, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.007787859998643398, + "learning_rate": 0.0012511546299672355, + "loss": 3.9498860836029053, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.00817918125540018, + "learning_rate": 0.0012498430112240246, + "loss": 4.003897666931152, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.009282548911869526, + "learning_rate": 0.001248531606113011, + "loss": 3.9621925354003906, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.009375610388815403, + "learning_rate": 0.001247220415690769, + "loss": 3.9977524280548096, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.009650378488004208, + "learning_rate": 0.001245909441013701, + "loss": 3.9882829189300537, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.009350919164717197, + "learning_rate": 0.0012445986831380366, + "loss": 4.019596099853516, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.008726303465664387, + "learning_rate": 0.0012432881431198289, + "loss": 3.9933621883392334, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.00829484686255455, + "learning_rate": 0.0012419778220149568, + "loss": 4.001237869262695, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.008744452148675919, + "learning_rate": 0.0012406677208791218, + "loss": 3.9323887825012207, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.008170492947101593, + "learning_rate": 0.0012393578407678496, + "loss": 3.925581455230713, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.007843422703444958, + "learning_rate": 0.001238048182736486, + "loss": 4.001023769378662, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.008914309553802013, + "learning_rate": 0.0012367387478401987, + "loss": 3.9804537296295166, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.009736751206219196, + "learning_rate": 0.0012354295371339774, + "loss": 3.9831767082214355, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.010061381384730339, + "learning_rate": 0.0012341205516726273, + "loss": 3.9462435245513916, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.010350692085921764, + "learning_rate": 0.0012328117925107765, + "loss": 4.012863636016846, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01039902213960886, + "learning_rate": 0.0012315032607028667, + "loss": 3.9811301231384277, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.011423371732234955, + "learning_rate": 0.0012301949573031595, + "loss": 3.9773542881011963, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.012449614703655243, + "learning_rate": 0.001228886883365731, + "loss": 3.969600200653076, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.010361429303884506, + "learning_rate": 0.0012275790399444726, + "loss": 4.035360336303711, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.009603082202374935, + "learning_rate": 0.0012262714280930904, + "loss": 3.9840846061706543, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.009183845482766628, + "learning_rate": 0.0012249640488651036, + "loss": 4.031542778015137, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.009694446809589863, + "learning_rate": 0.0012236569033138441, + "loss": 3.9935057163238525, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.0097816726192832, + "learning_rate": 0.0012223499924924545, + "loss": 3.9851176738739014, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.009413184598088264, + "learning_rate": 0.001221043317453891, + "loss": 3.9642205238342285, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.009960711002349854, + "learning_rate": 0.0012197368792509166, + "loss": 3.9373817443847656, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.010124670341610909, + "learning_rate": 0.0012184306789361059, + "loss": 3.958115577697754, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.011087656021118164, + "learning_rate": 0.00121712471756184, + "loss": 3.9607057571411133, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.009812800213694572, + "learning_rate": 0.001215818996180309, + "loss": 4.00722599029541, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.010305573232471943, + "learning_rate": 0.0012145135158435094, + "loss": 3.98728084564209, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.01019787136465311, + "learning_rate": 0.0012132082776032422, + "loss": 3.981172800064087, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.009260181337594986, + "learning_rate": 0.001211903282511115, + "loss": 4.034339904785156, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.00972998607903719, + "learning_rate": 0.001210598531618538, + "loss": 4.022885322570801, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.009853586554527283, + "learning_rate": 0.0012092940259767267, + "loss": 4.000235080718994, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.00961449183523655, + "learning_rate": 0.0012079897666366962, + "loss": 3.9951353073120117, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.010257176123559475, + "learning_rate": 0.0012066857546492657, + "loss": 3.9801371097564697, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.009443561546504498, + "learning_rate": 0.001205381991065054, + "loss": 3.9596502780914307, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.010389018803834915, + "learning_rate": 0.0012040784769344795, + "loss": 3.980527639389038, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.010322567075490952, + "learning_rate": 0.0012027752133077607, + "loss": 3.996258497238159, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.009987728670239449, + "learning_rate": 0.0012014722012349125, + "loss": 3.9493203163146973, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.421086311340332, + "eval_runtime": 52.5623, + "eval_samples_per_second": 46.459, + "eval_steps_per_second": 1.465, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.009405115619301796, + "learning_rate": 0.0012001694417657493, + "loss": 3.927506923675537, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.009974643588066101, + "learning_rate": 0.00119886693594988, + "loss": 3.9373550415039062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.011331096291542053, + "learning_rate": 0.0011975646848367098, + "loss": 3.935594081878662, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.010538448579609394, + "learning_rate": 0.0011962626894754405, + "loss": 3.911579132080078, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.00989572424441576, + "learning_rate": 0.0011949609509150645, + "loss": 3.92525577545166, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.010533696971833706, + "learning_rate": 0.0011936594702043704, + "loss": 3.898075580596924, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.01051455456763506, + "learning_rate": 0.0011923582483919365, + "loss": 3.93463134765625, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.009662075899541378, + "learning_rate": 0.0011910572865261348, + "loss": 3.898221731185913, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.008874096907675266, + "learning_rate": 0.0011897565856551257, + "loss": 3.926787853240967, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.008126807399094105, + "learning_rate": 0.0011884561468268606, + "loss": 3.9158413410186768, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.007915553636848927, + "learning_rate": 0.001187155971089081, + "loss": 3.8891496658325195, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.008312922902405262, + "learning_rate": 0.0011858560594893129, + "loss": 3.932219982147217, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.008892328478395939, + "learning_rate": 0.0011845564130748734, + "loss": 3.8866336345672607, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.009187526069581509, + "learning_rate": 0.0011832570328928628, + "loss": 3.9464707374572754, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.008615494705736637, + "learning_rate": 0.0011819579199901693, + "loss": 3.9396629333496094, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.008649308234453201, + "learning_rate": 0.0011806590754134636, + "loss": 3.9325060844421387, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.008581082336604595, + "learning_rate": 0.0011793605002092021, + "loss": 3.9200944900512695, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.008502699434757233, + "learning_rate": 0.0011780621954236238, + "loss": 3.9656100273132324, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.0083576999604702, + "learning_rate": 0.0011767641621027477, + "loss": 3.9130215644836426, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.008196269161999226, + "learning_rate": 0.0011754664012923785, + "loss": 3.942369222640991, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.008857307955622673, + "learning_rate": 0.0011741689140380956, + "loss": 3.909585952758789, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.008935349993407726, + "learning_rate": 0.0011728717013852635, + "loss": 3.9248366355895996, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.008821200579404831, + "learning_rate": 0.0011715747643790211, + "loss": 3.9589288234710693, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.00861942209303379, + "learning_rate": 0.0011702781040642885, + "loss": 3.9381728172302246, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.009127574972808361, + "learning_rate": 0.0011689817214857613, + "loss": 3.897829532623291, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.008008269593119621, + "learning_rate": 0.0011676856176879107, + "loss": 3.930056095123291, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.010070208460092545, + "learning_rate": 0.0011663897937149853, + "loss": 3.892270803451538, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.00903113093227148, + "learning_rate": 0.001165094250611006, + "loss": 3.9766440391540527, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.008872415870428085, + "learning_rate": 0.0011637989894197692, + "loss": 3.9559712409973145, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.009524816647171974, + "learning_rate": 0.0011625040111848427, + "loss": 3.9607882499694824, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.009836779907345772, + "learning_rate": 0.0011612093169495678, + "loss": 3.9629592895507812, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.009586540050804615, + "learning_rate": 0.0011599149077570561, + "loss": 3.95208740234375, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.010495901107788086, + "learning_rate": 0.0011586207846501896, + "loss": 3.9439730644226074, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.010346266441047192, + "learning_rate": 0.0011573269486716202, + "loss": 3.959299087524414, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.010767634026706219, + "learning_rate": 0.0011560334008637674, + "loss": 3.900784969329834, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.012122200801968575, + "learning_rate": 0.0011547401422688202, + "loss": 3.9588193893432617, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.012914645485579967, + "learning_rate": 0.0011534471739287331, + "loss": 3.9278407096862793, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.011785333044826984, + "learning_rate": 0.0011521544968852273, + "loss": 3.992793560028076, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.008455781266093254, + "learning_rate": 0.0011508621121797899, + "loss": 3.96525239944458, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.009655282832682133, + "learning_rate": 0.0011495700208536714, + "loss": 3.9346890449523926, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.010537004098296165, + "learning_rate": 0.0011482782239478871, + "loss": 3.9534668922424316, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.009438651613891125, + "learning_rate": 0.0011469867225032135, + "loss": 3.983487606048584, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.010586646385490894, + "learning_rate": 0.0011456955175601908, + "loss": 3.939680576324463, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.009082718752324581, + "learning_rate": 0.001144404610159119, + "loss": 3.9558515548706055, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.009833602234721184, + "learning_rate": 0.0011431140013400591, + "loss": 3.9539828300476074, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.009142170660197735, + "learning_rate": 0.0011418236921428313, + "loss": 3.9442849159240723, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.008603853173553944, + "learning_rate": 0.0011405336836070137, + "loss": 3.9526913166046143, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.008838302455842495, + "learning_rate": 0.0011392439767719446, + "loss": 3.973104953765869, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.008867006748914719, + "learning_rate": 0.0011379545726767156, + "loss": 3.9832305908203125, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.00971697736531496, + "learning_rate": 0.001136665472360178, + "loss": 3.9563937187194824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.00957175437361002, + "learning_rate": 0.001135376676860935, + "loss": 3.975194215774536, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.008548660203814507, + "learning_rate": 0.001134088187217347, + "loss": 3.9839985370635986, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.008556334301829338, + "learning_rate": 0.0011328000044675264, + "loss": 3.9225516319274902, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.008440867997705936, + "learning_rate": 0.0011315121296493385, + "loss": 3.945028066635132, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.009643333032727242, + "learning_rate": 0.0011302245638004017, + "loss": 3.943120002746582, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.008832835592329502, + "learning_rate": 0.001128937307958083, + "loss": 3.9895920753479004, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.008720474317669868, + "learning_rate": 0.0011276503631595036, + "loss": 3.941950798034668, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.008099926635622978, + "learning_rate": 0.001126363730441529, + "loss": 3.9626870155334473, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.008525107055902481, + "learning_rate": 0.0011250774108407778, + "loss": 3.950535774230957, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.008998272009193897, + "learning_rate": 0.0011237914053936135, + "loss": 3.9460678100585938, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.009387437254190445, + "learning_rate": 0.001122505715136148, + "loss": 3.926604986190796, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.007804017513990402, + "learning_rate": 0.0011212203411042386, + "loss": 3.9894230365753174, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.008564616553485394, + "learning_rate": 0.0011199352843334876, + "loss": 3.9339959621429443, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.008726440370082855, + "learning_rate": 0.0011186505458592427, + "loss": 3.9867987632751465, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.008074100129306316, + "learning_rate": 0.0011173661267165932, + "loss": 3.9492292404174805, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.007688041776418686, + "learning_rate": 0.0011160820279403744, + "loss": 3.9425454139709473, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.007808993570506573, + "learning_rate": 0.0011147982505651595, + "loss": 3.935114860534668, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.009786820039153099, + "learning_rate": 0.0011135147956252659, + "loss": 3.959758758544922, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.009487155824899673, + "learning_rate": 0.00111223166415475, + "loss": 3.989048480987549, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.008407804183661938, + "learning_rate": 0.0011109488571874073, + "loss": 3.932347297668457, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.007158380467444658, + "learning_rate": 0.0011096663757567723, + "loss": 3.9545154571533203, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.008840581402182579, + "learning_rate": 0.001108384220896117, + "loss": 3.965984344482422, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.409194469451904, + "eval_runtime": 51.5561, + "eval_samples_per_second": 47.366, + "eval_steps_per_second": 1.494, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.010702896863222122, + "learning_rate": 0.0011071023936384508, + "loss": 3.8964462280273438, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01290619745850563, + "learning_rate": 0.0011058208950165183, + "loss": 3.8775625228881836, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01443667896091938, + "learning_rate": 0.0011045397260628, + "loss": 3.8755288124084473, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.014411968179047108, + "learning_rate": 0.0011032588878095108, + "loss": 3.889918327331543, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.013656266033649445, + "learning_rate": 0.0011019783812885986, + "loss": 3.926997661590576, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.010661937296390533, + "learning_rate": 0.0011006982075317446, + "loss": 3.9226956367492676, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.008968768641352654, + "learning_rate": 0.0010994183675703615, + "loss": 3.8932647705078125, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.009992181323468685, + "learning_rate": 0.0010981388624355937, + "loss": 3.8920068740844727, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.011034460738301277, + "learning_rate": 0.0010968596931583148, + "loss": 3.909073829650879, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.011815821751952171, + "learning_rate": 0.0010955808607691284, + "loss": 3.9190006256103516, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.012317490763962269, + "learning_rate": 0.0010943023662983674, + "loss": 3.8790552616119385, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.010983958840370178, + "learning_rate": 0.0010930242107760903, + "loss": 3.8730852603912354, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.008948802947998047, + "learning_rate": 0.0010917463952320856, + "loss": 3.923659563064575, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.009568693116307259, + "learning_rate": 0.0010904689206958649, + "loss": 3.9100794792175293, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.011862963438034058, + "learning_rate": 0.0010891917881966672, + "loss": 3.908093214035034, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.010180995799601078, + "learning_rate": 0.0010879149987634545, + "loss": 3.900660514831543, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.009372464381158352, + "learning_rate": 0.0010866385534249131, + "loss": 3.8950703144073486, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.010695629753172398, + "learning_rate": 0.0010853624532094525, + "loss": 3.913530111312866, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.011817459017038345, + "learning_rate": 0.0010840866991452024, + "loss": 3.90268611907959, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.010686337016522884, + "learning_rate": 0.0010828112922600167, + "loss": 3.9141364097595215, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.009550466202199459, + "learning_rate": 0.0010815362335814655, + "loss": 3.968468189239502, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.009367208927869797, + "learning_rate": 0.0010802615241368423, + "loss": 3.8933968544006348, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.009899397380650043, + "learning_rate": 0.0010789871649531562, + "loss": 3.936708450317383, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.009181005880236626, + "learning_rate": 0.0010777131570571363, + "loss": 3.9370503425598145, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.00799303874373436, + "learning_rate": 0.0010764395014752274, + "loss": 3.911050319671631, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.008469315245747566, + "learning_rate": 0.0010751661992335908, + "loss": 3.9103240966796875, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.009814459830522537, + "learning_rate": 0.0010738932513581034, + "loss": 3.881843090057373, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.010244108736515045, + "learning_rate": 0.0010726206588743557, + "loss": 3.883141040802002, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.008898732252418995, + "learning_rate": 0.001071348422807653, + "loss": 3.952347755432129, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.00823870487511158, + "learning_rate": 0.0010700765441830121, + "loss": 3.932507038116455, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.009857657365500927, + "learning_rate": 0.001068805024025164, + "loss": 3.932932138442993, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.00929308496415615, + "learning_rate": 0.0010675338633585488, + "loss": 3.8971338272094727, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.008667961694300175, + "learning_rate": 0.0010662630632073172, + "loss": 3.916931629180908, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.009758287109434605, + "learning_rate": 0.001064992624595331, + "loss": 3.964726209640503, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.009429254569113255, + "learning_rate": 0.0010637225485461586, + "loss": 3.917572498321533, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.008303353562951088, + "learning_rate": 0.0010624528360830782, + "loss": 3.875241756439209, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.008729013614356518, + "learning_rate": 0.001061183488229073, + "loss": 3.9691600799560547, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.010239302180707455, + "learning_rate": 0.0010599145060068353, + "loss": 3.9053149223327637, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.00932721421122551, + "learning_rate": 0.0010586458904387596, + "loss": 3.8593955039978027, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.008196988143026829, + "learning_rate": 0.0010573776425469473, + "loss": 3.917983055114746, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.008049141615629196, + "learning_rate": 0.001056109763353203, + "loss": 3.9337785243988037, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.009256313554942608, + "learning_rate": 0.001054842253879034, + "loss": 3.937887668609619, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.009292179718613625, + "learning_rate": 0.0010535751151456495, + "loss": 3.9203450679779053, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.009142161346971989, + "learning_rate": 0.0010523083481739602, + "loss": 3.9027457237243652, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.009498288854956627, + "learning_rate": 0.0010510419539845777, + "loss": 3.917316436767578, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.00849096104502678, + "learning_rate": 0.001049775933597813, + "loss": 3.9229235649108887, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.008343606255948544, + "learning_rate": 0.0010485102880336747, + "loss": 3.9138073921203613, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.010342134162783623, + "learning_rate": 0.0010472450183118727, + "loss": 3.8932929039001465, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.010298306122422218, + "learning_rate": 0.00104598012545181, + "loss": 3.9112677574157715, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.00827560480684042, + "learning_rate": 0.001044715610472589, + "loss": 3.894556760787964, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.008993102237582207, + "learning_rate": 0.0010434514743930064, + "loss": 3.905391216278076, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.008632538840174675, + "learning_rate": 0.0010421877182315536, + "loss": 3.896714210510254, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.008542424067854881, + "learning_rate": 0.0010409243430064162, + "loss": 3.9145569801330566, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.008970450609922409, + "learning_rate": 0.0010396613497354725, + "loss": 3.9170360565185547, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.008744142949581146, + "learning_rate": 0.0010383987394362942, + "loss": 3.927412509918213, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.008229847066104412, + "learning_rate": 0.0010371365131261424, + "loss": 3.928352117538452, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.00818693358451128, + "learning_rate": 0.0010358746718219717, + "loss": 3.9219343662261963, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.008848588913679123, + "learning_rate": 0.001034613216540423, + "loss": 3.9346699714660645, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.00834780652076006, + "learning_rate": 0.0010333521482978295, + "loss": 3.9144492149353027, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.00940527394413948, + "learning_rate": 0.0010320914681102098, + "loss": 3.951573371887207, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.008884521201252937, + "learning_rate": 0.0010308311769932725, + "loss": 3.899625778198242, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.007863261736929417, + "learning_rate": 0.0010295712759624107, + "loss": 3.9321727752685547, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.008651153184473515, + "learning_rate": 0.0010283117660327035, + "loss": 3.9030094146728516, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.00822074618190527, + "learning_rate": 0.0010270526482189163, + "loss": 3.939466953277588, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.008701153099536896, + "learning_rate": 0.0010257939235354964, + "loss": 3.9099626541137695, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.008418620564043522, + "learning_rate": 0.0010245355929965774, + "loss": 3.914116382598877, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.008339700289070606, + "learning_rate": 0.001023277657615971, + "loss": 3.9111878871917725, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.008304647170007229, + "learning_rate": 0.0010220201184071747, + "loss": 3.950974464416504, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.008436229079961777, + "learning_rate": 0.001020762976383365, + "loss": 3.9631457328796387, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.0074487063102424145, + "learning_rate": 0.001019506232557398, + "loss": 3.902984619140625, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.008736515417695045, + "learning_rate": 0.0010182498879418096, + "loss": 3.9268949031829834, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.009666468016803265, + "learning_rate": 0.0010169939435488145, + "loss": 3.9274301528930664, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.395970821380615, + "eval_runtime": 53.7848, + "eval_samples_per_second": 45.403, + "eval_steps_per_second": 1.432, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.009094545617699623, + "learning_rate": 0.001015738400390304, + "loss": 3.866425037384033, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.009273733012378216, + "learning_rate": 0.0010144832594778462, + "loss": 3.8569841384887695, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.009205271489918232, + "learning_rate": 0.0010132285218226855, + "loss": 3.8981378078460693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.009963045828044415, + "learning_rate": 0.0010119741884357424, + "loss": 3.881542444229126, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.008536554872989655, + "learning_rate": 0.0010107202603276096, + "loss": 3.851253032684326, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.01074561569839716, + "learning_rate": 0.001009466738508555, + "loss": 3.8731184005737305, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.010809902101755142, + "learning_rate": 0.0010082136239885177, + "loss": 3.863889217376709, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.008289352990686893, + "learning_rate": 0.0010069609177771103, + "loss": 3.839066743850708, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.00792208593338728, + "learning_rate": 0.001005708620883615, + "loss": 3.8634843826293945, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.008381598629057407, + "learning_rate": 0.0010044567343169845, + "loss": 3.868283987045288, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.009784617461264133, + "learning_rate": 0.0010032052590858425, + "loss": 3.853074550628662, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.009462198242545128, + "learning_rate": 0.0010019541961984782, + "loss": 3.889979600906372, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.010207819752395153, + "learning_rate": 0.001000703546662852, + "loss": 3.8636741638183594, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.009123551659286022, + "learning_rate": 0.0009994533114865887, + "loss": 3.8578555583953857, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.009006140753626823, + "learning_rate": 0.0009982034916769806, + "loss": 3.8828015327453613, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.0091114342212677, + "learning_rate": 0.0009969540882409846, + "loss": 3.880486011505127, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.008851367980241776, + "learning_rate": 0.0009957051021852227, + "loss": 3.8618125915527344, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.008196741342544556, + "learning_rate": 0.0009944565345159806, + "loss": 3.8314175605773926, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.00863513071089983, + "learning_rate": 0.000993208386239206, + "loss": 3.856212615966797, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.008624681271612644, + "learning_rate": 0.0009919606583605107, + "loss": 3.868927001953125, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.009208842180669308, + "learning_rate": 0.0009907133518851647, + "loss": 3.9072189331054688, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.010200913064181805, + "learning_rate": 0.0009894664678181027, + "loss": 3.875272750854492, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.009892710484564304, + "learning_rate": 0.000988220007163914, + "loss": 3.888686418533325, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.009873585775494576, + "learning_rate": 0.000986973970926851, + "loss": 3.8607118129730225, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.009322221390902996, + "learning_rate": 0.0009857283601108228, + "loss": 3.8852052688598633, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.008141528815031052, + "learning_rate": 0.0009844831757193947, + "loss": 3.888213872909546, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.008191613480448723, + "learning_rate": 0.00098323841875579, + "loss": 3.9073967933654785, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.008419081568717957, + "learning_rate": 0.0009819940902228864, + "loss": 3.879223346710205, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.009037869982421398, + "learning_rate": 0.0009807501911232171, + "loss": 3.883303165435791, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.00854078121483326, + "learning_rate": 0.0009795067224589687, + "loss": 3.926767587661743, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.008590771816670895, + "learning_rate": 0.0009782636852319825, + "loss": 3.913999557495117, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.00859815627336502, + "learning_rate": 0.0009770210804437498, + "loss": 3.8828063011169434, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.008970260620117188, + "learning_rate": 0.0009757789090954157, + "loss": 3.9174866676330566, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.00877175759524107, + "learning_rate": 0.0009745371721877754, + "loss": 3.8879401683807373, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.008541764691472054, + "learning_rate": 0.0009732958707212731, + "loss": 3.8717427253723145, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.008466241881251335, + "learning_rate": 0.0009720550056960039, + "loss": 3.8944485187530518, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.00808491837233305, + "learning_rate": 0.0009708145781117096, + "loss": 3.8617026805877686, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.00842207670211792, + "learning_rate": 0.0009695745889677805, + "loss": 3.887824535369873, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.00830599945038557, + "learning_rate": 0.0009683350392632533, + "loss": 3.8612070083618164, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.008417914621531963, + "learning_rate": 0.0009670959299968111, + "loss": 3.880140781402588, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.008542059920728207, + "learning_rate": 0.0009658572621667822, + "loss": 3.8951029777526855, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.008822236210107803, + "learning_rate": 0.000964619036771138, + "loss": 3.8962583541870117, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.009151184931397438, + "learning_rate": 0.0009633812548074952, + "loss": 3.882500648498535, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.007803036365658045, + "learning_rate": 0.0009621439172731115, + "loss": 3.8740782737731934, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.009350850246846676, + "learning_rate": 0.0009609070251648884, + "loss": 3.8627419471740723, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.008319144137203693, + "learning_rate": 0.0009596705794793667, + "loss": 3.9084486961364746, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.009198528714478016, + "learning_rate": 0.0009584345812127282, + "loss": 3.92141056060791, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.010480938479304314, + "learning_rate": 0.0009571990313607958, + "loss": 3.8916103839874268, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.01045288611203432, + "learning_rate": 0.0009559639309190278, + "loss": 3.8863210678100586, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.009121391922235489, + "learning_rate": 0.0009547292808825239, + "loss": 3.9167044162750244, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.01093223039060831, + "learning_rate": 0.0009534950822460184, + "loss": 3.9046711921691895, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.01214868575334549, + "learning_rate": 0.0009522613360038835, + "loss": 3.9126768112182617, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.01055280864238739, + "learning_rate": 0.0009510280431501255, + "loss": 3.9214353561401367, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.009021313861012459, + "learning_rate": 0.0009497952046783867, + "loss": 3.8795580863952637, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.00863752793520689, + "learning_rate": 0.0009485628215819428, + "loss": 3.893232822418213, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.009327626787126064, + "learning_rate": 0.0009473308948537021, + "loss": 3.917377471923828, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.01003898587077856, + "learning_rate": 0.0009460994254862066, + "loss": 3.8818931579589844, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.00892532430589199, + "learning_rate": 0.0009448684144716278, + "loss": 3.8761978149414062, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.008396463468670845, + "learning_rate": 0.0009436378628017702, + "loss": 3.8665928840637207, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.00784370768815279, + "learning_rate": 0.0009424077714680662, + "loss": 3.8850138187408447, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.00809321366250515, + "learning_rate": 0.0009411781414615784, + "loss": 3.917078733444214, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.008685288950800896, + "learning_rate": 0.000939948973772998, + "loss": 3.8865013122558594, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.007291820365935564, + "learning_rate": 0.0009387202693926425, + "loss": 3.8967666625976562, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.008459319360554218, + "learning_rate": 0.0009374920293104573, + "loss": 3.9043235778808594, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.007924137637019157, + "learning_rate": 0.0009362642545160123, + "loss": 3.9216856956481934, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.007874754257500172, + "learning_rate": 0.0009350369459985052, + "loss": 3.879725933074951, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.008643723092973232, + "learning_rate": 0.0009338101047467544, + "loss": 3.9303886890411377, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.008451704867184162, + "learning_rate": 0.000932583731749205, + "loss": 3.9352827072143555, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.008137754164636135, + "learning_rate": 0.0009313578279939236, + "loss": 3.930800437927246, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.009277567267417908, + "learning_rate": 0.0009301323944685979, + "loss": 3.9194300174713135, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.009428970515727997, + "learning_rate": 0.0009289074321605383, + "loss": 3.874298572540283, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.0089727733284235, + "learning_rate": 0.0009276829420566743, + "loss": 3.893770694732666, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.388528823852539, + "eval_runtime": 51.5859, + "eval_samples_per_second": 47.338, + "eval_steps_per_second": 1.493, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.009470812045037746, + "learning_rate": 0.000926458925143556, + "loss": 3.8514318466186523, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.011485518887639046, + "learning_rate": 0.0009252353824073511, + "loss": 3.857734203338623, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.010095939971506596, + "learning_rate": 0.0009240123148338457, + "loss": 3.819117546081543, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.010103728622198105, + "learning_rate": 0.0009227897234084444, + "loss": 3.842660427093506, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.01148147787898779, + "learning_rate": 0.0009215676091161661, + "loss": 3.821089029312134, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.012331553734838963, + "learning_rate": 0.0009203459729416467, + "loss": 3.8194339275360107, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.011274794116616249, + "learning_rate": 0.0009191248158691358, + "loss": 3.8839240074157715, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.009603365324437618, + "learning_rate": 0.0009179041388824986, + "loss": 3.806617259979248, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.010824795812368393, + "learning_rate": 0.0009166839429652112, + "loss": 3.823228597640991, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.01014280878007412, + "learning_rate": 0.0009154642291003637, + "loss": 3.860605239868164, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.009052124805748463, + "learning_rate": 0.0009142449982706591, + "loss": 3.84226131439209, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.009465130977332592, + "learning_rate": 0.0009130262514584074, + "loss": 3.8674912452697754, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.009991783648729324, + "learning_rate": 0.0009118079896455326, + "loss": 3.8775739669799805, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.009288253262639046, + "learning_rate": 0.0009105902138135652, + "loss": 3.7978098392486572, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.009964323602616787, + "learning_rate": 0.0009093729249436459, + "loss": 3.8362321853637695, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.010977157391607761, + "learning_rate": 0.0009081561240165218, + "loss": 3.857304096221924, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.009241319261491299, + "learning_rate": 0.0009069398120125479, + "loss": 3.8659310340881348, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.009211038239300251, + "learning_rate": 0.0009057239899116844, + "loss": 3.8764538764953613, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.009649050422012806, + "learning_rate": 0.000904508658693497, + "loss": 3.8702778816223145, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.00954380165785551, + "learning_rate": 0.0009032938193371573, + "loss": 3.8273327350616455, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.008831270970404148, + "learning_rate": 0.0009020794728214378, + "loss": 3.821664810180664, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.008367286063730717, + "learning_rate": 0.000900865620124717, + "loss": 3.8586034774780273, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.008805174380540848, + "learning_rate": 0.0008996522622249732, + "loss": 3.811522960662842, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.00873226672410965, + "learning_rate": 0.0008984394000997873, + "loss": 3.8538150787353516, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.008930203504860401, + "learning_rate": 0.0008972270347263402, + "loss": 3.8448486328125, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.009300434030592442, + "learning_rate": 0.0008960151670814125, + "loss": 3.842113494873047, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.009598663076758385, + "learning_rate": 0.0008948037981413855, + "loss": 3.886852741241455, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.008400370366871357, + "learning_rate": 0.0008935929288822357, + "loss": 3.852076292037964, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.00891344528645277, + "learning_rate": 0.0008923825602795391, + "loss": 3.8480606079101562, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.00869396049529314, + "learning_rate": 0.0008911726933084678, + "loss": 3.8164706230163574, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.007914266549050808, + "learning_rate": 0.0008899633289437904, + "loss": 3.8768272399902344, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.008398533798754215, + "learning_rate": 0.0008887544681598687, + "loss": 3.8625502586364746, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.00887146033346653, + "learning_rate": 0.0008875461119306601, + "loss": 3.8934593200683594, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.008642833679914474, + "learning_rate": 0.0008863382612297172, + "loss": 3.8391976356506348, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.008099788799881935, + "learning_rate": 0.0008851309170301816, + "loss": 3.827965497970581, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.008166342042386532, + "learning_rate": 0.0008839240803047894, + "loss": 3.8999152183532715, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.009095479734241962, + "learning_rate": 0.0008827177520258671, + "loss": 3.876927614212036, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.009246554225683212, + "learning_rate": 0.0008815119331653321, + "loss": 3.8692450523376465, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.007059223018586636, + "learning_rate": 0.0008803066246946898, + "loss": 3.869400978088379, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.008499744348227978, + "learning_rate": 0.0008791018275850363, + "loss": 3.8303189277648926, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.007667839527130127, + "learning_rate": 0.0008778975428070543, + "loss": 3.827998638153076, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.00765089550986886, + "learning_rate": 0.0008766937713310145, + "loss": 3.8554482460021973, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.00862938817590475, + "learning_rate": 0.0008754905141267749, + "loss": 3.905458927154541, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.008733712136745453, + "learning_rate": 0.0008742877721637758, + "loss": 3.8564133644104004, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.0071738120168447495, + "learning_rate": 0.0008730855464110471, + "loss": 3.878807544708252, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.008418241515755653, + "learning_rate": 0.0008718838378371989, + "loss": 3.8399720191955566, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.008526502177119255, + "learning_rate": 0.0008706826474104263, + "loss": 3.847656488418579, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.00732462527230382, + "learning_rate": 0.0008694819760985069, + "loss": 3.8839492797851562, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.008424196392297745, + "learning_rate": 0.0008682818248688003, + "loss": 3.861239433288574, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.008984955959022045, + "learning_rate": 0.0008670821946882465, + "loss": 3.856529712677002, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.009004034101963043, + "learning_rate": 0.000865883086523365, + "loss": 3.8514528274536133, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.009097260423004627, + "learning_rate": 0.0008646845013402563, + "loss": 3.8890581130981445, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.009244748391211033, + "learning_rate": 0.0008634864401045988, + "loss": 3.872745990753174, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.00788815412670374, + "learning_rate": 0.0008622889037816488, + "loss": 3.8777575492858887, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.008943610824644566, + "learning_rate": 0.0008610918933362396, + "loss": 3.8825736045837402, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.008154108189046383, + "learning_rate": 0.0008598954097327809, + "loss": 3.877514362335205, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.008689315989613533, + "learning_rate": 0.0008586994539352587, + "loss": 3.854720115661621, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.00904505793005228, + "learning_rate": 0.0008575040269072315, + "loss": 3.911463499069214, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.007773210760205984, + "learning_rate": 0.0008563091296118342, + "loss": 3.875370740890503, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.008437697775661945, + "learning_rate": 0.0008551147630117737, + "loss": 3.8457469940185547, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.009780957363545895, + "learning_rate": 0.0008539209280693296, + "loss": 3.890479564666748, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.007929213345050812, + "learning_rate": 0.0008527276257463532, + "loss": 3.876030445098877, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.009121386334300041, + "learning_rate": 0.0008515348570042669, + "loss": 3.9252734184265137, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.00981531199067831, + "learning_rate": 0.000850342622804063, + "loss": 3.8997983932495117, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.009150844067335129, + "learning_rate": 0.0008491509241063025, + "loss": 3.85115385055542, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.00992978923022747, + "learning_rate": 0.0008479597618711154, + "loss": 3.9138102531433105, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.010906689800322056, + "learning_rate": 0.0008467691370582001, + "loss": 3.892329692840576, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.008644519373774529, + "learning_rate": 0.0008455790506268213, + "loss": 3.9071311950683594, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.009030954912304878, + "learning_rate": 0.0008443895035358111, + "loss": 3.867384910583496, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.010806960985064507, + "learning_rate": 0.0008432004967435638, + "loss": 3.877983331680298, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.011148087680339813, + "learning_rate": 0.0008420120312080436, + "loss": 3.8795924186706543, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01046178862452507, + "learning_rate": 0.0008408241078867731, + "loss": 3.887709617614746, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.383835792541504, + "eval_runtime": 53.0778, + "eval_samples_per_second": 46.008, + "eval_steps_per_second": 1.451, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.009721867740154266, + "learning_rate": 0.0008396367277368421, + "loss": 3.8261327743530273, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.01091420091688633, + "learning_rate": 0.000838449891714901, + "loss": 3.7837276458740234, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.010423852130770683, + "learning_rate": 0.0008372636007771623, + "loss": 3.796450138092041, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.011928882449865341, + "learning_rate": 0.0008360778558793996, + "loss": 3.794717311859131, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.010779444128274918, + "learning_rate": 0.000834892657976945, + "loss": 3.8477983474731445, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.009554282762110233, + "learning_rate": 0.0008337080080246915, + "loss": 3.793790817260742, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.009768152609467506, + "learning_rate": 0.0008325239069770906, + "loss": 3.860384225845337, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.010801832191646099, + "learning_rate": 0.0008313403557881513, + "loss": 3.816796064376831, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.01046004518866539, + "learning_rate": 0.0008301573554114378, + "loss": 3.8064301013946533, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.009319447912275791, + "learning_rate": 0.0008289749068000744, + "loss": 3.832408905029297, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01011407095938921, + "learning_rate": 0.000827793010906737, + "loss": 3.8276009559631348, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.010393989272415638, + "learning_rate": 0.0008266116686836582, + "loss": 3.83754825592041, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.00883004255592823, + "learning_rate": 0.0008254308810826244, + "loss": 3.8315513134002686, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.008549787104129791, + "learning_rate": 0.0008242506490549745, + "loss": 3.792919158935547, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.009668390266597271, + "learning_rate": 0.0008230709735516006, + "loss": 3.8110954761505127, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.009008699096739292, + "learning_rate": 0.0008218918555229445, + "loss": 3.8598508834838867, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.008660725317895412, + "learning_rate": 0.0008207132959190026, + "loss": 3.839632511138916, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.00827972125262022, + "learning_rate": 0.0008195352956893172, + "loss": 3.8428683280944824, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.008117233403027058, + "learning_rate": 0.0008183578557829825, + "loss": 3.8434956073760986, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.008314455859363079, + "learning_rate": 0.0008171809771486405, + "loss": 3.8079686164855957, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.00865850131958723, + "learning_rate": 0.0008160046607344811, + "loss": 3.846999406814575, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0096864253282547, + "learning_rate": 0.0008148289074882416, + "loss": 3.8227243423461914, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.00840236246585846, + "learning_rate": 0.0008136537183572045, + "loss": 3.8400344848632812, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.007451791316270828, + "learning_rate": 0.0008124790942881984, + "loss": 3.8318095207214355, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.00858070608228445, + "learning_rate": 0.000811305036227597, + "loss": 3.8232922554016113, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.00765061192214489, + "learning_rate": 0.0008101315451213178, + "loss": 3.81624174118042, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.009304090403020382, + "learning_rate": 0.0008089586219148213, + "loss": 3.8715004920959473, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.009136030450463295, + "learning_rate": 0.0008077862675531103, + "loss": 3.825037956237793, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.007460939697921276, + "learning_rate": 0.0008066144829807303, + "loss": 3.811749219894409, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.007510480936616659, + "learning_rate": 0.0008054432691417659, + "loss": 3.8313679695129395, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.008825462311506271, + "learning_rate": 0.0008042726269798433, + "loss": 3.83119535446167, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.009002318605780602, + "learning_rate": 0.0008031025574381281, + "loss": 3.8542468547821045, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.009266122244298458, + "learning_rate": 0.0008019330614593237, + "loss": 3.8328065872192383, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.008935115300118923, + "learning_rate": 0.0008007641399856727, + "loss": 3.8198719024658203, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.008021850138902664, + "learning_rate": 0.0007995957939589526, + "loss": 3.828382968902588, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.008035171777009964, + "learning_rate": 0.0007984280243204805, + "loss": 3.879868507385254, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.008553579449653625, + "learning_rate": 0.0007972608320111059, + "loss": 3.842162609100342, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.008016746491193771, + "learning_rate": 0.000796094217971215, + "loss": 3.85674786567688, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.007781418971717358, + "learning_rate": 0.0007949281831407276, + "loss": 3.7968859672546387, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.008552255108952522, + "learning_rate": 0.0007937627284590971, + "loss": 3.8305506706237793, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.007390565704554319, + "learning_rate": 0.0007925978548653096, + "loss": 3.8228843212127686, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.007611850742250681, + "learning_rate": 0.0007914335632978812, + "loss": 3.868957996368408, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.008188549429178238, + "learning_rate": 0.0007902698546948628, + "loss": 3.836078405380249, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.008198797702789307, + "learning_rate": 0.0007891067299938318, + "loss": 3.8637044429779053, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.008353965356945992, + "learning_rate": 0.0007879441901318977, + "loss": 3.833843946456909, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.0086967209354043, + "learning_rate": 0.0007867822360456963, + "loss": 3.850536823272705, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.009151408448815346, + "learning_rate": 0.0007856208686713949, + "loss": 3.8453407287597656, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.008010942488908768, + "learning_rate": 0.0007844600889446858, + "loss": 3.8679513931274414, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.007813206873834133, + "learning_rate": 0.0007832998978007873, + "loss": 3.860609531402588, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.007759213913232088, + "learning_rate": 0.0007821402961744448, + "loss": 3.803863048553467, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.008045864291489124, + "learning_rate": 0.0007809812849999293, + "loss": 3.858161449432373, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.008212368935346603, + "learning_rate": 0.0007798228652110347, + "loss": 3.8311336040496826, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.007982932962477207, + "learning_rate": 0.0007786650377410774, + "loss": 3.870339870452881, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.0071402136236429214, + "learning_rate": 0.0007775078035229002, + "loss": 3.7922263145446777, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.008564750663936138, + "learning_rate": 0.0007763511634888656, + "loss": 3.83981990814209, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.008078585378825665, + "learning_rate": 0.0007751951185708562, + "loss": 3.852938652038574, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.00801754929125309, + "learning_rate": 0.0007740396697002774, + "loss": 3.8434009552001953, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.008973612450063229, + "learning_rate": 0.0007728848178080535, + "loss": 3.8238563537597656, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.010199865326285362, + "learning_rate": 0.0007717305638246281, + "loss": 3.8891735076904297, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.009620471857488155, + "learning_rate": 0.0007705769086799617, + "loss": 3.8611769676208496, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.007981563918292522, + "learning_rate": 0.0007694238533035337, + "loss": 3.8496928215026855, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.008856567554175854, + "learning_rate": 0.0007682713986243415, + "loss": 3.8366098403930664, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.008323253132402897, + "learning_rate": 0.0007671195455708952, + "loss": 3.8224244117736816, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.00814893003553152, + "learning_rate": 0.0007659682950712228, + "loss": 3.8611154556274414, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.009146672673523426, + "learning_rate": 0.0007648176480528659, + "loss": 3.8258843421936035, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.007573423441499472, + "learning_rate": 0.0007636676054428805, + "loss": 3.888007640838623, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.00813984964042902, + "learning_rate": 0.0007625181681678343, + "loss": 3.860306739807129, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.007936778478324413, + "learning_rate": 0.0007613693371538078, + "loss": 3.8808555603027344, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.008434669114649296, + "learning_rate": 0.0007602211133263956, + "loss": 3.854154348373413, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.007857564836740494, + "learning_rate": 0.0007590734976106985, + "loss": 3.8240737915039062, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.007989327423274517, + "learning_rate": 0.0007579264909313315, + "loss": 3.850971221923828, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.008888393640518188, + "learning_rate": 0.0007567800942124152, + "loss": 3.862267255783081, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.38022518157959, + "eval_runtime": 51.4686, + "eval_samples_per_second": 47.446, + "eval_steps_per_second": 1.496, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.00952722318470478, + "learning_rate": 0.0007556343083775832, + "loss": 3.7705495357513428, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01099323108792305, + "learning_rate": 0.0007544891343499723, + "loss": 3.80812406539917, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.010298571549355984, + "learning_rate": 0.0007533445730522295, + "loss": 3.782794237136841, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.008139589801430702, + "learning_rate": 0.000752200625406507, + "loss": 3.7877392768859863, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.009031252935528755, + "learning_rate": 0.0007510572923344628, + "loss": 3.8181564807891846, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.00942459236830473, + "learning_rate": 0.0007499145747572603, + "loss": 3.807425022125244, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.00922947097569704, + "learning_rate": 0.0007487724735955648, + "loss": 3.7923264503479004, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.009551790542900562, + "learning_rate": 0.0007476309897695485, + "loss": 3.790083885192871, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.008996761403977871, + "learning_rate": 0.0007464901241988831, + "loss": 3.836491107940674, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.009766182862222195, + "learning_rate": 0.0007453498778027433, + "loss": 3.7829647064208984, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.007708575110882521, + "learning_rate": 0.0007442102514998058, + "loss": 3.803400993347168, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.008606902323663235, + "learning_rate": 0.0007430712462082464, + "loss": 3.810825824737549, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.009462927468121052, + "learning_rate": 0.0007419328628457416, + "loss": 3.784748077392578, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.010283366777002811, + "learning_rate": 0.0007407951023294655, + "loss": 3.744312286376953, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.008952794596552849, + "learning_rate": 0.0007396579655760917, + "loss": 3.8098015785217285, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.008446207270026207, + "learning_rate": 0.0007385214535017907, + "loss": 3.8306760787963867, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.007988285273313522, + "learning_rate": 0.0007373855670222302, + "loss": 3.7976696491241455, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.007586437277495861, + "learning_rate": 0.0007362503070525718, + "loss": 3.785891056060791, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.00775605533272028, + "learning_rate": 0.0007351156745074757, + "loss": 3.8176379203796387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.007727691903710365, + "learning_rate": 0.0007339816703010948, + "loss": 3.791743278503418, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.008311517536640167, + "learning_rate": 0.0007328482953470747, + "loss": 3.818769693374634, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.007610874716192484, + "learning_rate": 0.0007317155505585556, + "loss": 3.76772141456604, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.0078107016161084175, + "learning_rate": 0.0007305834368481699, + "loss": 3.822758197784424, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.006906864698976278, + "learning_rate": 0.0007294519551280412, + "loss": 3.832834005355835, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.008230660110712051, + "learning_rate": 0.0007283211063097829, + "loss": 3.85087513923645, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.008201118558645248, + "learning_rate": 0.0007271908913044997, + "loss": 3.7777247428894043, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.0070146857760846615, + "learning_rate": 0.000726061311022787, + "loss": 3.837517738342285, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.0076036700047552586, + "learning_rate": 0.0007249323663747254, + "loss": 3.8358592987060547, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.007238952908664942, + "learning_rate": 0.0007238040582698858, + "loss": 3.828089952468872, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.00825269054621458, + "learning_rate": 0.0007226763876173259, + "loss": 3.806016445159912, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.008368568494915962, + "learning_rate": 0.0007215493553255898, + "loss": 3.8163394927978516, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.00728471577167511, + "learning_rate": 0.0007204229623027063, + "loss": 3.835597038269043, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.00784830842167139, + "learning_rate": 0.0007192972094561897, + "loss": 3.8050971031188965, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.007714141625910997, + "learning_rate": 0.0007181720976930404, + "loss": 3.8031606674194336, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.009128103032708168, + "learning_rate": 0.000717047627919739, + "loss": 3.77107572555542, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.008033272810280323, + "learning_rate": 0.000715923801042251, + "loss": 3.8615059852600098, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.007732565049082041, + "learning_rate": 0.0007148006179660235, + "loss": 3.8277132511138916, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.007382966112345457, + "learning_rate": 0.0007136780795959853, + "loss": 3.803431987762451, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.007517020218074322, + "learning_rate": 0.0007125561868365442, + "loss": 3.799734354019165, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.008041434921324253, + "learning_rate": 0.0007114349405915895, + "loss": 3.7848258018493652, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.007875649258494377, + "learning_rate": 0.0007103143417644891, + "loss": 3.8136563301086426, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.008213640190660954, + "learning_rate": 0.000709194391258089, + "loss": 3.8518495559692383, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.00801054760813713, + "learning_rate": 0.0007080750899747139, + "loss": 3.7956271171569824, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.008064007386565208, + "learning_rate": 0.0007069564388161631, + "loss": 3.854595184326172, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.008300521411001682, + "learning_rate": 0.0007058384386837154, + "loss": 3.8203625679016113, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.008640496991574764, + "learning_rate": 0.000704721090478122, + "loss": 3.793574810028076, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.007684200536459684, + "learning_rate": 0.0007036043950996105, + "loss": 3.8564438819885254, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.008431905880570412, + "learning_rate": 0.0007024883534478823, + "loss": 3.8271119594573975, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.008137186989188194, + "learning_rate": 0.0007013729664221122, + "loss": 3.8219337463378906, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.0076744635589420795, + "learning_rate": 0.0007002582349209474, + "loss": 3.8353049755096436, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.008388740941882133, + "learning_rate": 0.0006991441598425065, + "loss": 3.795104503631592, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.007856911979615688, + "learning_rate": 0.0006980307420843798, + "loss": 3.818411111831665, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.008048887364566326, + "learning_rate": 0.000696917982543628, + "loss": 3.86599063873291, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.007949979975819588, + "learning_rate": 0.0006958058821167812, + "loss": 3.8026742935180664, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.008162821643054485, + "learning_rate": 0.0006946944416998389, + "loss": 3.833751678466797, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.008190765976905823, + "learning_rate": 0.0006935836621882682, + "loss": 3.8164236545562744, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.009714310988783836, + "learning_rate": 0.000692473544477005, + "loss": 3.8367366790771484, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.008075987920165062, + "learning_rate": 0.0006913640894604498, + "loss": 3.8108773231506348, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.009075167588889599, + "learning_rate": 0.0006902552980324713, + "loss": 3.8444485664367676, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.008678440935909748, + "learning_rate": 0.0006891471710864022, + "loss": 3.8384461402893066, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.007705411408096552, + "learning_rate": 0.0006880397095150414, + "loss": 3.825443744659424, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.00782099086791277, + "learning_rate": 0.0006869329142106488, + "loss": 3.8331687450408936, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.007542501203715801, + "learning_rate": 0.0006858267860649509, + "loss": 3.7905991077423096, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.008315361104905605, + "learning_rate": 0.0006847213259691356, + "loss": 3.861527442932129, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.008529976941645145, + "learning_rate": 0.0006836165348138508, + "loss": 3.8332972526550293, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.008117211051285267, + "learning_rate": 0.0006825124134892072, + "loss": 3.7961089611053467, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.008409497328102589, + "learning_rate": 0.0006814089628847755, + "loss": 3.812218189239502, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.007592010777443647, + "learning_rate": 0.0006803061838895864, + "loss": 3.801304340362549, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.007820053026080132, + "learning_rate": 0.0006792040773921283, + "loss": 3.8517117500305176, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.0073044621385633945, + "learning_rate": 0.0006781026442803479, + "loss": 3.8253560066223145, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.007546546868979931, + "learning_rate": 0.0006770018854416522, + "loss": 3.800802230834961, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.008878474123775959, + "learning_rate": 0.0006759018017629007, + "loss": 3.8133859634399414, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.375829219818115, + "eval_runtime": 51.3756, + "eval_samples_per_second": 47.532, + "eval_steps_per_second": 1.499, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.00814585480839014, + "learning_rate": 0.0006748023941304115, + "loss": 3.7970919609069824, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.00824262760579586, + "learning_rate": 0.0006737036634299577, + "loss": 3.750131607055664, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.008675593882799149, + "learning_rate": 0.000672605610546767, + "loss": 3.791456699371338, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.009052547626197338, + "learning_rate": 0.00067150823636552, + "loss": 3.7762060165405273, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.00818016566336155, + "learning_rate": 0.0006704115417703514, + "loss": 3.750631332397461, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.00800115056335926, + "learning_rate": 0.0006693155276448489, + "loss": 3.811519145965576, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.009032433852553368, + "learning_rate": 0.0006682201948720509, + "loss": 3.7805609703063965, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.0077886078506708145, + "learning_rate": 0.0006671255443344474, + "loss": 3.749366521835327, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.008603204973042011, + "learning_rate": 0.0006660315769139774, + "loss": 3.7686312198638916, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.008361323736608028, + "learning_rate": 0.0006649382934920329, + "loss": 3.740269660949707, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.00948982872068882, + "learning_rate": 0.0006638456949494507, + "loss": 3.7874045372009277, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.008958354592323303, + "learning_rate": 0.0006627537821665185, + "loss": 3.78987455368042, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.008702470920979977, + "learning_rate": 0.0006616625560229704, + "loss": 3.751061201095581, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.008424391970038414, + "learning_rate": 0.000660572017397988, + "loss": 3.80816650390625, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.009022071026265621, + "learning_rate": 0.0006594821671701989, + "loss": 3.726067304611206, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.007974586449563503, + "learning_rate": 0.0006583930062176739, + "loss": 3.78564453125, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.007887622341513634, + "learning_rate": 0.0006573045354179329, + "loss": 3.7993059158325195, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.007782666943967342, + "learning_rate": 0.0006562167556479352, + "loss": 3.7440507411956787, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.007923880591988564, + "learning_rate": 0.000655129667784086, + "loss": 3.7920784950256348, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.007850701920688152, + "learning_rate": 0.0006540432727022325, + "loss": 3.7833633422851562, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.007855751551687717, + "learning_rate": 0.0006529575712776633, + "loss": 3.812349319458008, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.007832377217710018, + "learning_rate": 0.0006518725643851091, + "loss": 3.7815146446228027, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.006997724529355764, + "learning_rate": 0.0006507882528987395, + "loss": 3.795346975326538, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.007433670572936535, + "learning_rate": 0.0006497046376921648, + "loss": 3.841017246246338, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.0075767990201711655, + "learning_rate": 0.0006486217196384344, + "loss": 3.7913427352905273, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.007551983464509249, + "learning_rate": 0.0006475394996100359, + "loss": 3.765105962753296, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.007235546596348286, + "learning_rate": 0.0006464579784788943, + "loss": 3.7668676376342773, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.007872921414673328, + "learning_rate": 0.0006453771571163716, + "loss": 3.770536422729492, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.007598782889544964, + "learning_rate": 0.0006442970363932667, + "loss": 3.796818733215332, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.007902889512479305, + "learning_rate": 0.0006432176171798124, + "loss": 3.824531078338623, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.008964902721345425, + "learning_rate": 0.0006421389003456776, + "loss": 3.781778335571289, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.009936625137925148, + "learning_rate": 0.000641060886759965, + "loss": 3.750248670578003, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.008381214924156666, + "learning_rate": 0.0006399835772912106, + "loss": 3.797954797744751, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.007752922363579273, + "learning_rate": 0.000638906972807384, + "loss": 3.789350748062134, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.009501751512289047, + "learning_rate": 0.0006378310741758841, + "loss": 3.791079521179199, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.00936005637049675, + "learning_rate": 0.0006367558822635453, + "loss": 3.78501033782959, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.008370982483029366, + "learning_rate": 0.0006356813979366284, + "loss": 3.792647361755371, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.00782831758260727, + "learning_rate": 0.000634607622060827, + "loss": 3.772569179534912, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.00830658059567213, + "learning_rate": 0.0006335345555012627, + "loss": 3.7787656784057617, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.009015833027660847, + "learning_rate": 0.000632462199122486, + "loss": 3.787418842315674, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.008993860334157944, + "learning_rate": 0.0006313905537884752, + "loss": 3.818718433380127, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.007954884320497513, + "learning_rate": 0.0006303196203626344, + "loss": 3.8100008964538574, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.007742907386273146, + "learning_rate": 0.0006292493997077973, + "loss": 3.7786598205566406, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.008748230524361134, + "learning_rate": 0.0006281798926862195, + "loss": 3.798635482788086, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.008306438103318214, + "learning_rate": 0.000627111100159585, + "loss": 3.827213764190674, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.008027350530028343, + "learning_rate": 0.0006260430229889987, + "loss": 3.7682337760925293, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.007779776118695736, + "learning_rate": 0.0006249756620349931, + "loss": 3.7515475749969482, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.008748390711843967, + "learning_rate": 0.000623909018157521, + "loss": 3.7963085174560547, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.008587505668401718, + "learning_rate": 0.0006228430922159577, + "loss": 3.7767529487609863, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.007172323763370514, + "learning_rate": 0.0006217778850691007, + "loss": 3.79768705368042, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.007224083412438631, + "learning_rate": 0.000620713397575168, + "loss": 3.785390853881836, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.008209837600588799, + "learning_rate": 0.0006196496305917987, + "loss": 3.809828281402588, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.007760563865303993, + "learning_rate": 0.0006185865849760491, + "loss": 3.838716506958008, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.007362260017544031, + "learning_rate": 0.0006175242615843977, + "loss": 3.7840137481689453, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.008719335310161114, + "learning_rate": 0.0006164626612727381, + "loss": 3.8228893280029297, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.00856029987335205, + "learning_rate": 0.0006154017848963826, + "loss": 3.80448317527771, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.008078506216406822, + "learning_rate": 0.0006143416333100606, + "loss": 3.7881453037261963, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.007957885973155499, + "learning_rate": 0.0006132822073679164, + "loss": 3.814870834350586, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.009362339973449707, + "learning_rate": 0.0006122235079235114, + "loss": 3.798335075378418, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.008272632025182247, + "learning_rate": 0.0006111655358298194, + "loss": 3.8212976455688477, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.007548908703029156, + "learning_rate": 0.0006101082919392301, + "loss": 3.8312830924987793, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.007543765474110842, + "learning_rate": 0.0006090517771035457, + "loss": 3.8302183151245117, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.007520351093262434, + "learning_rate": 0.000607995992173981, + "loss": 3.7898802757263184, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.007168850861489773, + "learning_rate": 0.000606940938001163, + "loss": 3.8078577518463135, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.007420001085847616, + "learning_rate": 0.0006058866154351296, + "loss": 3.799571990966797, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.007621012162417173, + "learning_rate": 0.0006048330253253303, + "loss": 3.833843946456909, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.007661828305572271, + "learning_rate": 0.0006037801685206222, + "loss": 3.820556640625, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.007657836657017469, + "learning_rate": 0.0006027280458692738, + "loss": 3.798994541168213, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.007722760550677776, + "learning_rate": 0.0006016766582189613, + "loss": 3.8589251041412354, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.00789247453212738, + "learning_rate": 0.0006006260064167685, + "loss": 3.8673787117004395, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.009137817658483982, + "learning_rate": 0.0005995760913091874, + "loss": 3.801938056945801, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.008374894969165325, + "learning_rate": 0.0005985269137421141, + "loss": 3.820045232772827, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.3730878829956055, + "eval_runtime": 48.1607, + "eval_samples_per_second": 50.705, + "eval_steps_per_second": 1.599, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.009492472745478153, + "learning_rate": 0.0005974784745608536, + "loss": 3.738617420196533, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.010017318651080132, + "learning_rate": 0.0005964307746101135, + "loss": 3.7551231384277344, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.008958659134805202, + "learning_rate": 0.000595383814734007, + "loss": 3.7739763259887695, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.008034365251660347, + "learning_rate": 0.0005943375957760505, + "loss": 3.7678627967834473, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.008489780128002167, + "learning_rate": 0.000593292118579164, + "loss": 3.720778465270996, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.009107839316129684, + "learning_rate": 0.0005922473839856699, + "loss": 3.7224440574645996, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.009295578114688396, + "learning_rate": 0.0005912033928372906, + "loss": 3.7695257663726807, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.007896862924098969, + "learning_rate": 0.0005901601459751532, + "loss": 3.742567300796509, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.008227395825088024, + "learning_rate": 0.0005891176442397808, + "loss": 3.7547202110290527, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.009326796047389507, + "learning_rate": 0.0005880758884710992, + "loss": 3.724808931350708, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.007786506786942482, + "learning_rate": 0.0005870348795084319, + "loss": 3.7403640747070312, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.00821041688323021, + "learning_rate": 0.0005859946181905009, + "loss": 3.770012617111206, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.009544595144689083, + "learning_rate": 0.000584955105355427, + "loss": 3.7378525733947754, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.008973932825028896, + "learning_rate": 0.0005839163418407254, + "loss": 3.7184295654296875, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.009023688733577728, + "learning_rate": 0.00058287832848331, + "loss": 3.7388229370117188, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.008151066489517689, + "learning_rate": 0.0005818410661194889, + "loss": 3.7632875442504883, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.008495030924677849, + "learning_rate": 0.0005808045555849668, + "loss": 3.8114736080169678, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.008739210665225983, + "learning_rate": 0.0005797687977148394, + "loss": 3.7647829055786133, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.008062886074185371, + "learning_rate": 0.0005787337933435999, + "loss": 3.782172441482544, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.007615398149937391, + "learning_rate": 0.0005776995433051326, + "loss": 3.7564642429351807, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.007938290014863014, + "learning_rate": 0.0005766660484327133, + "loss": 3.756208896636963, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.008517666719853878, + "learning_rate": 0.0005756333095590102, + "loss": 3.76729416847229, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.007426336407661438, + "learning_rate": 0.0005746013275160827, + "loss": 3.74001407623291, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.007664753124117851, + "learning_rate": 0.0005735701031353804, + "loss": 3.7819876670837402, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.007903067395091057, + "learning_rate": 0.0005725396372477414, + "loss": 3.781219959259033, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.007251975126564503, + "learning_rate": 0.0005715099306833931, + "loss": 3.763620615005493, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.007422176189720631, + "learning_rate": 0.0005704809842719533, + "loss": 3.775099754333496, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.009161600843071938, + "learning_rate": 0.0005694527988424239, + "loss": 3.792268991470337, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.008330529555678368, + "learning_rate": 0.0005684253752231958, + "loss": 3.7967824935913086, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.0073860278353095055, + "learning_rate": 0.0005673987142420458, + "loss": 3.76505970954895, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.008462096564471722, + "learning_rate": 0.0005663728167261363, + "loss": 3.740450143814087, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.00838128849864006, + "learning_rate": 0.0005653476835020138, + "loss": 3.7863235473632812, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.007381014991551638, + "learning_rate": 0.0005643233153956092, + "loss": 3.772179126739502, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.007363761775195599, + "learning_rate": 0.0005632997132322392, + "loss": 3.7793867588043213, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.008684969507157803, + "learning_rate": 0.0005622768778366001, + "loss": 3.7736339569091797, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.007934059016406536, + "learning_rate": 0.0005612548100327722, + "loss": 3.8040053844451904, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.007136524189263582, + "learning_rate": 0.0005602335106442172, + "loss": 3.7589468955993652, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.00789916142821312, + "learning_rate": 0.0005592129804937779, + "loss": 3.7953195571899414, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.007919179275631905, + "learning_rate": 0.0005581932204036761, + "loss": 3.7539429664611816, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.007734223268926144, + "learning_rate": 0.0005571742311955147, + "loss": 3.7620325088500977, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.007005508057773113, + "learning_rate": 0.0005561560136902749, + "loss": 3.762538433074951, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.008137146942317486, + "learning_rate": 0.0005551385687083161, + "loss": 3.773789405822754, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.007310070563107729, + "learning_rate": 0.0005541218970693757, + "loss": 3.8077168464660645, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.007335536181926727, + "learning_rate": 0.0005531059995925661, + "loss": 3.7761921882629395, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.008022893220186234, + "learning_rate": 0.00055209087709638, + "loss": 3.7769880294799805, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.008027128875255585, + "learning_rate": 0.0005510765303986814, + "loss": 3.7709951400756836, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.00697955721989274, + "learning_rate": 0.0005500629603167118, + "loss": 3.7855210304260254, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.007455077022314072, + "learning_rate": 0.0005490501676670861, + "loss": 3.7850632667541504, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.0069657545536756516, + "learning_rate": 0.0005480381532657932, + "loss": 3.742671251296997, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.007609451189637184, + "learning_rate": 0.0005470269179281953, + "loss": 3.793771266937256, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.007437967229634523, + "learning_rate": 0.0005460164624690255, + "loss": 3.778846263885498, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.007123028859496117, + "learning_rate": 0.00054500678770239, + "loss": 3.768375873565674, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.008176218718290329, + "learning_rate": 0.0005439978944417654, + "loss": 3.82069730758667, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.006943001411855221, + "learning_rate": 0.000542989783499999, + "loss": 3.7731552124023438, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.0072205523028969765, + "learning_rate": 0.0005419824556893069, + "loss": 3.7724058628082275, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.007656400091946125, + "learning_rate": 0.0005409759118212758, + "loss": 3.780989646911621, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.006972154602408409, + "learning_rate": 0.0005399701527068599, + "loss": 3.7971787452697754, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.007674331311136484, + "learning_rate": 0.0005389651791563806, + "loss": 3.798673391342163, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.006952045019716024, + "learning_rate": 0.0005379609919795268, + "loss": 3.7982821464538574, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.007882075384259224, + "learning_rate": 0.0005369575919853545, + "loss": 3.825962781906128, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.007143386639654636, + "learning_rate": 0.0005359549799822854, + "loss": 3.826598644256592, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.007587207015603781, + "learning_rate": 0.0005349531567781041, + "loss": 3.80623197555542, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.007677781395614147, + "learning_rate": 0.0005339521231799632, + "loss": 3.7987256050109863, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.007372343447059393, + "learning_rate": 0.0005329518799943773, + "loss": 3.7733969688415527, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.00861304346472025, + "learning_rate": 0.0005319524280272231, + "loss": 3.7903685569763184, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.00816237274557352, + "learning_rate": 0.0005309537680837418, + "loss": 3.77058744430542, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.007624274119734764, + "learning_rate": 0.0005299559009685356, + "loss": 3.787982940673828, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.007416521664708853, + "learning_rate": 0.0005289588274855683, + "loss": 3.7991762161254883, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.007461614441126585, + "learning_rate": 0.0005279625484381632, + "loss": 3.8036863803863525, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.007766335736960173, + "learning_rate": 0.0005269670646290043, + "loss": 3.809243679046631, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.0073394919745624065, + "learning_rate": 0.0005259723768601365, + "loss": 3.756606101989746, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.008771199733018875, + "learning_rate": 0.0005249784859329602, + "loss": 3.773951768875122, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.370637893676758, + "eval_runtime": 49.4904, + "eval_samples_per_second": 49.343, + "eval_steps_per_second": 1.556, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.008164486847817898, + "learning_rate": 0.0005239853926482358, + "loss": 3.7619595527648926, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.007980132475495338, + "learning_rate": 0.0005229930978060806, + "loss": 3.7398204803466797, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.0077056242153048515, + "learning_rate": 0.0005220016022059692, + "loss": 3.699580669403076, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.0074698119424283504, + "learning_rate": 0.0005210109066467308, + "loss": 3.7299928665161133, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.008384468965232372, + "learning_rate": 0.0005200210119265512, + "loss": 3.704989433288574, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.008066141977906227, + "learning_rate": 0.0005190319188429709, + "loss": 3.751697540283203, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.007213297300040722, + "learning_rate": 0.0005180436281928841, + "loss": 3.740692377090454, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.007343464996665716, + "learning_rate": 0.0005170561407725396, + "loss": 3.7060189247131348, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.00783725269138813, + "learning_rate": 0.0005160694573775363, + "loss": 3.7414426803588867, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.007640232797712088, + "learning_rate": 0.0005150835788028294, + "loss": 3.7451486587524414, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.008129071444272995, + "learning_rate": 0.0005140985058427218, + "loss": 3.734839916229248, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.006920623127371073, + "learning_rate": 0.0005131142392908697, + "loss": 3.7138614654541016, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.00798740889877081, + "learning_rate": 0.0005121307799402787, + "loss": 3.76496958732605, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.006689135450869799, + "learning_rate": 0.0005111481285833045, + "loss": 3.7707948684692383, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.007544311694800854, + "learning_rate": 0.0005101662860116516, + "loss": 3.7393479347229004, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.007487910334020853, + "learning_rate": 0.0005091852530163717, + "loss": 3.7486305236816406, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.00772349676117301, + "learning_rate": 0.0005082050303878673, + "loss": 3.753629207611084, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.007883663289248943, + "learning_rate": 0.0005072256189158844, + "loss": 3.7497611045837402, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.008153906092047691, + "learning_rate": 0.0005062470193895178, + "loss": 3.7459521293640137, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.007083641365170479, + "learning_rate": 0.0005052692325972075, + "loss": 3.725599765777588, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.007102683652192354, + "learning_rate": 0.0005042922593267385, + "loss": 3.7724428176879883, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.008104706183075905, + "learning_rate": 0.0005033161003652409, + "loss": 3.7386724948883057, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.0073765418492257595, + "learning_rate": 0.0005023407564991878, + "loss": 3.7478132247924805, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.007110628765076399, + "learning_rate": 0.0005013662285143963, + "loss": 3.771925687789917, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.007323848083615303, + "learning_rate": 0.000500392517196026, + "loss": 3.7460904121398926, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.007830807007849216, + "learning_rate": 0.0004994196233285786, + "loss": 3.751222610473633, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.007710007485002279, + "learning_rate": 0.0004984475476958973, + "loss": 3.739858627319336, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.007293973118066788, + "learning_rate": 0.0004974762910811655, + "loss": 3.7410120964050293, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.0073824538849294186, + "learning_rate": 0.0004965058542669078, + "loss": 3.711397647857666, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.007156423293054104, + "learning_rate": 0.0004955362380349866, + "loss": 3.7881593704223633, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.007705818396061659, + "learning_rate": 0.0004945674431666047, + "loss": 3.7639269828796387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.0075300210155546665, + "learning_rate": 0.0004935994704423026, + "loss": 3.7562999725341797, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.007137116510421038, + "learning_rate": 0.000492632320641959, + "loss": 3.7659149169921875, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.007225630804896355, + "learning_rate": 0.0004916659945447878, + "loss": 3.7625155448913574, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.008027992211282253, + "learning_rate": 0.0004907004929293408, + "loss": 3.746478796005249, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.007582477759569883, + "learning_rate": 0.0004897358165735066, + "loss": 3.759408950805664, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.007575645577162504, + "learning_rate": 0.000488771966254506, + "loss": 3.7430644035339355, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.008404700085520744, + "learning_rate": 0.0004878089427488963, + "loss": 3.7727198600769043, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.007118664216250181, + "learning_rate": 0.00048684674683256815, + "loss": 3.747460126876831, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.007220778148621321, + "learning_rate": 0.00048588537928074584, + "loss": 3.746413230895996, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.0076194945722818375, + "learning_rate": 0.0004849248408679849, + "loss": 3.7808125019073486, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.007285538129508495, + "learning_rate": 0.00048396513236817387, + "loss": 3.7750301361083984, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.007698678877204657, + "learning_rate": 0.00048300625455453383, + "loss": 3.7474594116210938, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.007687424309551716, + "learning_rate": 0.0004820482081996141, + "loss": 3.7929720878601074, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.007363605313003063, + "learning_rate": 0.0004810909940752959, + "loss": 3.75911021232605, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.007791243493556976, + "learning_rate": 0.00048013461295278836, + "loss": 3.7567291259765625, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.007375436834990978, + "learning_rate": 0.0004791790656026323, + "loss": 3.7784414291381836, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.007330593653023243, + "learning_rate": 0.0004782243527946934, + "loss": 3.7533810138702393, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.006926248781383038, + "learning_rate": 0.00047727047529816694, + "loss": 3.7795910835266113, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.007102136500179768, + "learning_rate": 0.00047631743388157496, + "loss": 3.812549114227295, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.006676106248050928, + "learning_rate": 0.00047536522931276555, + "loss": 3.7578792572021484, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.006926767062395811, + "learning_rate": 0.0004744138623589134, + "loss": 3.782416820526123, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.006894777063280344, + "learning_rate": 0.000473463333786516, + "loss": 3.7840042114257812, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.007393381092697382, + "learning_rate": 0.00047251364436139923, + "loss": 3.7532780170440674, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.006510896608233452, + "learning_rate": 0.0004715647948487092, + "loss": 3.7618460655212402, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.007283828221261501, + "learning_rate": 0.0004706167860129171, + "loss": 3.768498420715332, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.00655644154176116, + "learning_rate": 0.0004696696186178168, + "loss": 3.7724053859710693, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.007338619790971279, + "learning_rate": 0.0004687232934265239, + "loss": 3.7814595699310303, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.006839970126748085, + "learning_rate": 0.0004677778112014758, + "loss": 3.771808624267578, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.006776005961000919, + "learning_rate": 0.0004668331727044297, + "loss": 3.786924123764038, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.007011357229202986, + "learning_rate": 0.00046588937869646415, + "loss": 3.787672996520996, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.006482086144387722, + "learning_rate": 0.0004649464299379771, + "loss": 3.7794556617736816, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.0069654518738389015, + "learning_rate": 0.000464004327188685, + "loss": 3.7520360946655273, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.006526473443955183, + "learning_rate": 0.0004630630712076235, + "loss": 3.734144926071167, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.006822322495281696, + "learning_rate": 0.00046212266275314504, + "loss": 3.7614707946777344, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.006827171426266432, + "learning_rate": 0.00046118310258292045, + "loss": 3.755218029022217, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.00706718722358346, + "learning_rate": 0.00046024439145393546, + "loss": 3.7397823333740234, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.00689381267875433, + "learning_rate": 0.00045930653012249314, + "loss": 3.758922576904297, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.007069882471114397, + "learning_rate": 0.0004583695193442113, + "loss": 3.759767532348633, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.006889182608574629, + "learning_rate": 0.0004574333598740228, + "loss": 3.7555460929870605, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.006866258103400469, + "learning_rate": 0.00045649805246617475, + "loss": 3.762068271636963, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.008696220815181732, + "learning_rate": 0.000455563597874226, + "loss": 3.7110466957092285, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.369833469390869, + "eval_runtime": 48.7852, + "eval_samples_per_second": 50.056, + "eval_steps_per_second": 1.578, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.0077184769324958324, + "learning_rate": 0.0004546299968510516, + "loss": 3.704669713973999, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.00846019759774208, + "learning_rate": 0.0004536972501488358, + "loss": 3.701132297515869, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.008113945834338665, + "learning_rate": 0.00045276535851907613, + "loss": 3.699582099914551, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.0072370911948382854, + "learning_rate": 0.0004518343227125807, + "loss": 3.698126792907715, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.00817187875509262, + "learning_rate": 0.0004509041434794685, + "loss": 3.738393783569336, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.007495569530874491, + "learning_rate": 0.00044997482156916845, + "loss": 3.727083683013916, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.007855897769331932, + "learning_rate": 0.0004490463577304171, + "loss": 3.716137170791626, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.007817867211997509, + "learning_rate": 0.0004481187527112628, + "loss": 3.750791072845459, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.007712380029261112, + "learning_rate": 0.00044719200725905887, + "loss": 3.720554828643799, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.00859921332448721, + "learning_rate": 0.00044626612212046816, + "loss": 3.742496967315674, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.007692081853747368, + "learning_rate": 0.0004453410980414582, + "loss": 3.703803300857544, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.007734883576631546, + "learning_rate": 0.00044441693576730565, + "loss": 3.7371301651000977, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.009022004902362823, + "learning_rate": 0.0004434936360425916, + "loss": 3.727294683456421, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.008058780804276466, + "learning_rate": 0.00044257119961120067, + "loss": 3.7583885192871094, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.00698013324290514, + "learning_rate": 0.00044164962721632376, + "loss": 3.7312440872192383, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.007999536581337452, + "learning_rate": 0.00044072891960045523, + "loss": 3.6972708702087402, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.008125107735395432, + "learning_rate": 0.0004398090775053929, + "loss": 3.7123398780822754, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.006736223120242357, + "learning_rate": 0.0004388901016722357, + "loss": 3.749678134918213, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.009154478088021278, + "learning_rate": 0.000437971992841387, + "loss": 3.722728967666626, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.008070871233940125, + "learning_rate": 0.00043705475175255087, + "loss": 3.7297017574310303, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.007498566992580891, + "learning_rate": 0.0004361383791447307, + "loss": 3.7286131381988525, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.007642887998372316, + "learning_rate": 0.00043522287575623183, + "loss": 3.750898838043213, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.007800443097949028, + "learning_rate": 0.0004343082423246588, + "loss": 3.719512462615967, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.007393732666969299, + "learning_rate": 0.0004333944795869159, + "loss": 3.7136006355285645, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.008559555746614933, + "learning_rate": 0.0004324815882792042, + "loss": 3.7087020874023438, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.007379850372672081, + "learning_rate": 0.0004315695691370241, + "loss": 3.7638165950775146, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.007374506443738937, + "learning_rate": 0.00043065842289517303, + "loss": 3.7208821773529053, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.00755207147449255, + "learning_rate": 0.0004297481502877451, + "loss": 3.738490581512451, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.007752610836178064, + "learning_rate": 0.0004288387520481306, + "loss": 3.7401833534240723, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.006907653529196978, + "learning_rate": 0.0004279302289090151, + "loss": 3.7754998207092285, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.007771312724798918, + "learning_rate": 0.0004270225816023797, + "loss": 3.73626446723938, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.007395982276648283, + "learning_rate": 0.0004261158108594987, + "loss": 3.7159175872802734, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.007196040358394384, + "learning_rate": 0.0004252099174109411, + "loss": 3.750246047973633, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.008241173811256886, + "learning_rate": 0.0004243049019865692, + "loss": 3.70859694480896, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.0072947596199810505, + "learning_rate": 0.00042340076531553755, + "loss": 3.7385268211364746, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.00782464537769556, + "learning_rate": 0.00042249750812629297, + "loss": 3.75885272026062, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.007813232019543648, + "learning_rate": 0.0004215951311465722, + "loss": 3.71810245513916, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.006900500506162643, + "learning_rate": 0.00042069363510340604, + "loss": 3.7365174293518066, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.007812316995114088, + "learning_rate": 0.0004197930207231119, + "loss": 3.774794101715088, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.007018622010946274, + "learning_rate": 0.00041889328873129933, + "loss": 3.7112040519714355, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.006895892322063446, + "learning_rate": 0.0004179944398528659, + "loss": 3.7497243881225586, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.0071940235793590546, + "learning_rate": 0.0004170964748119979, + "loss": 3.7566123008728027, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.006776201073080301, + "learning_rate": 0.0004161993943321702, + "loss": 3.7683968544006348, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.006886148825287819, + "learning_rate": 0.0004153031991361428, + "loss": 3.759763240814209, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.006466342136263847, + "learning_rate": 0.0004144078899459659, + "loss": 3.736006498336792, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.007055724039673805, + "learning_rate": 0.0004135134674829725, + "loss": 3.70645809173584, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.006948428228497505, + "learning_rate": 0.00041261993246778303, + "loss": 3.7365002632141113, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.006491438951343298, + "learning_rate": 0.00041172728562030255, + "loss": 3.735001564025879, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.006851755082607269, + "learning_rate": 0.0004108355276597209, + "loss": 3.720893383026123, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.007266475819051266, + "learning_rate": 0.00040994465930451114, + "loss": 3.7746100425720215, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.007112737279385328, + "learning_rate": 0.00040905468127242926, + "loss": 3.7441720962524414, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.006514255423098803, + "learning_rate": 0.000408165594280515, + "loss": 3.7333898544311523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.007172788958996534, + "learning_rate": 0.0004072773990450896, + "loss": 3.7675576210021973, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.00684684282168746, + "learning_rate": 0.0004063900962817564, + "loss": 3.7141308784484863, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.006578145083039999, + "learning_rate": 0.00040550368670539777, + "loss": 3.7256317138671875, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.007047492545098066, + "learning_rate": 0.00040461817103017926, + "loss": 3.7618556022644043, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.007036720868200064, + "learning_rate": 0.00040373354996954487, + "loss": 3.7397944927215576, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.00745626725256443, + "learning_rate": 0.00040284982423621714, + "loss": 3.7831153869628906, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.007454817183315754, + "learning_rate": 0.00040196699454219797, + "loss": 3.7409911155700684, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.0075650583021342754, + "learning_rate": 0.00040108506159876756, + "loss": 3.7074005603790283, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.006908242590725422, + "learning_rate": 0.0004002040261164838, + "loss": 3.767744779586792, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.0072857895866036415, + "learning_rate": 0.0003993238888051801, + "loss": 3.7349488735198975, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.0072942771948874, + "learning_rate": 0.00039844465037396716, + "loss": 3.75968861579895, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.0071496241725981236, + "learning_rate": 0.00039756631153123296, + "loss": 3.8237380981445312, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.006803249940276146, + "learning_rate": 0.00039668887298463783, + "loss": 3.7468650341033936, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.007433352991938591, + "learning_rate": 0.00039581233544111867, + "loss": 3.783236503601074, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.006633558310568333, + "learning_rate": 0.0003949366996068858, + "loss": 3.75020694732666, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.007415154483169317, + "learning_rate": 0.00039406196618742376, + "loss": 3.7545979022979736, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.006997843272984028, + "learning_rate": 0.0003931881358874886, + "loss": 3.7404396533966064, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.007417330518364906, + "learning_rate": 0.0003923152094111095, + "loss": 3.7575364112854004, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.006936701480299234, + "learning_rate": 0.0003914431874615889, + "loss": 3.771845817565918, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.008293208666145802, + "learning_rate": 0.00039057207074149795, + "loss": 3.746669292449951, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.370476245880127, + "eval_runtime": 50.9368, + "eval_samples_per_second": 47.942, + "eval_steps_per_second": 1.512, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.0077544888481497765, + "learning_rate": 0.0003897018599526799, + "loss": 3.6915597915649414, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.008674043230712414, + "learning_rate": 0.000388832555796248, + "loss": 3.67698073387146, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.00724575063213706, + "learning_rate": 0.00038796415897258523, + "loss": 3.6838502883911133, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.00849415734410286, + "learning_rate": 0.00038709667018134246, + "loss": 3.69064998626709, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.00769044877961278, + "learning_rate": 0.0003862300901214403, + "loss": 3.670502185821533, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.007846124470233917, + "learning_rate": 0.00038536441949106647, + "loss": 3.767771005630493, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.007555077783763409, + "learning_rate": 0.000384499658987676, + "loss": 3.7001633644104004, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.007834251970052719, + "learning_rate": 0.0003836358093079915, + "loss": 3.704331874847412, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.007608097977936268, + "learning_rate": 0.00038277287114799974, + "loss": 3.7108519077301025, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.007420446258038282, + "learning_rate": 0.0003819108452029561, + "loss": 3.697843313217163, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.008091001771390438, + "learning_rate": 0.000381049732167378, + "loss": 3.7380661964416504, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.007265515625476837, + "learning_rate": 0.00038018953273504907, + "loss": 3.680748462677002, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.007735357619822025, + "learning_rate": 0.00037933024759901636, + "loss": 3.6648077964782715, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.008188895881175995, + "learning_rate": 0.0003784718774515907, + "loss": 3.706609010696411, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.00816027820110321, + "learning_rate": 0.00037761442298434554, + "loss": 3.7368297576904297, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.006931936834007502, + "learning_rate": 0.00037675788488811566, + "loss": 3.7121026515960693, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.007615855894982815, + "learning_rate": 0.00037590226385299897, + "loss": 3.7379372119903564, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.0073505607433617115, + "learning_rate": 0.0003750475605683535, + "loss": 3.7330894470214844, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.007151979021728039, + "learning_rate": 0.00037419377572279897, + "loss": 3.713761568069458, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.007310053799301386, + "learning_rate": 0.00037334091000421326, + "loss": 3.705451011657715, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.0072968751192092896, + "learning_rate": 0.0003724889640997357, + "loss": 3.7550907135009766, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.007197601720690727, + "learning_rate": 0.00037163793869576404, + "loss": 3.7346463203430176, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.007823911495506763, + "learning_rate": 0.00037078783447795307, + "loss": 3.7257699966430664, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.007118315435945988, + "learning_rate": 0.0003699386521312171, + "loss": 3.7014496326446533, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.0069230347871780396, + "learning_rate": 0.00036909039233972666, + "loss": 3.6709237098693848, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.007836844772100449, + "learning_rate": 0.00036824305578691, + "loss": 3.695725440979004, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.007123398594558239, + "learning_rate": 0.00036739664315544956, + "loss": 3.7278594970703125, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.007792143151164055, + "learning_rate": 0.00036655115512728586, + "loss": 3.7291927337646484, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.007245976477861404, + "learning_rate": 0.00036570659238361345, + "loss": 3.7183046340942383, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.007778054103255272, + "learning_rate": 0.0003648629556048804, + "loss": 3.7485761642456055, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.007539471611380577, + "learning_rate": 0.00036402024547079007, + "loss": 3.7273335456848145, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.007554007228463888, + "learning_rate": 0.0003631784626602987, + "loss": 3.684359073638916, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.007862667553126812, + "learning_rate": 0.00036233760785161606, + "loss": 3.7223072052001953, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.00869659148156643, + "learning_rate": 0.0003614976817222028, + "loss": 3.7475168704986572, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.00696196174249053, + "learning_rate": 0.00036065868494877226, + "loss": 3.7773914337158203, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.008952745236456394, + "learning_rate": 0.00035982061820729005, + "loss": 3.6761155128479004, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.008341601118445396, + "learning_rate": 0.00035898348217297034, + "loss": 3.746553421020508, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.007268146146088839, + "learning_rate": 0.00035814727752027885, + "loss": 3.6921701431274414, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.007401647046208382, + "learning_rate": 0.00035731200492293064, + "loss": 3.722433090209961, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.006969883106648922, + "learning_rate": 0.00035647766505389006, + "loss": 3.7449448108673096, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.007616426330059767, + "learning_rate": 0.00035564425858536895, + "loss": 3.725113868713379, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.006883042864501476, + "learning_rate": 0.00035481178618882814, + "loss": 3.717106342315674, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.007106961216777563, + "learning_rate": 0.00035398024853497574, + "loss": 3.74045467376709, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.007402827963232994, + "learning_rate": 0.00035314964629376664, + "loss": 3.7288095951080322, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.006770852487534285, + "learning_rate": 0.00035231998013440227, + "loss": 3.7663369178771973, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.007622968405485153, + "learning_rate": 0.0003514912507253287, + "loss": 3.747084140777588, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.006876434665173292, + "learning_rate": 0.00035066345873423975, + "loss": 3.7591025829315186, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.007071215659379959, + "learning_rate": 0.00034983660482807116, + "loss": 3.7583885192871094, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.007130554411560297, + "learning_rate": 0.00034901068967300465, + "loss": 3.6950902938842773, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.006981188431382179, + "learning_rate": 0.00034818571393446524, + "loss": 3.7390527725219727, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.0067793820053339005, + "learning_rate": 0.0003473616782771209, + "loss": 3.703673839569092, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.00674026133492589, + "learning_rate": 0.0003465385833648827, + "loss": 3.6988182067871094, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.006553220096975565, + "learning_rate": 0.0003457164298609021, + "loss": 3.7367048263549805, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.007079749833792448, + "learning_rate": 0.000344895218427575, + "loss": 3.773226022720337, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.006091540213674307, + "learning_rate": 0.00034407494972653547, + "loss": 3.747929096221924, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.006783034652471542, + "learning_rate": 0.00034325562441865935, + "loss": 3.7154481410980225, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.006683558225631714, + "learning_rate": 0.0003424372431640625, + "loss": 3.751605987548828, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.0066695502027869225, + "learning_rate": 0.0003416198066220997, + "loss": 3.7561278343200684, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.006601935718208551, + "learning_rate": 0.0003408033154513655, + "loss": 3.754878520965576, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.006911450996994972, + "learning_rate": 0.00033998777030969085, + "loss": 3.69533109664917, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.0070345234125852585, + "learning_rate": 0.0003391731718541465, + "loss": 3.7328338623046875, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.007238153368234634, + "learning_rate": 0.00033835952074103964, + "loss": 3.7370123863220215, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.007295276038348675, + "learning_rate": 0.0003375468176259142, + "loss": 3.7123947143554688, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.006471444386988878, + "learning_rate": 0.0003367350631635506, + "loss": 3.73347806930542, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.007136075291782618, + "learning_rate": 0.00033592425800796447, + "loss": 3.7337799072265625, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.0065771411173045635, + "learning_rate": 0.00033511440281240725, + "loss": 3.733506441116333, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.006566311698406935, + "learning_rate": 0.0003343054982293643, + "loss": 3.7325632572174072, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.006389300804585218, + "learning_rate": 0.00033349754491055535, + "loss": 3.7027206420898438, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.006425539031624794, + "learning_rate": 0.00033269054350693386, + "loss": 3.751466989517212, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.006735894829034805, + "learning_rate": 0.0003318844946686865, + "loss": 3.731363296508789, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.0061791036278009415, + "learning_rate": 0.00033107939904523264, + "loss": 3.7592201232910156, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.007923831231892109, + "learning_rate": 0.0003302752572852219, + "loss": 3.7247159481048584, + "step": 2880 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.2278453445328896e+18, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-2880/training_args.bin b/runs/baseline/checkpoint-2880/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-2880/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-3240/chat_template.jinja b/runs/baseline/checkpoint-3240/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-3240/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-3240/config.json b/runs/baseline/checkpoint-3240/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-3240/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-3240/generation_config.json b/runs/baseline/checkpoint-3240/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-3240/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-3240/model.safetensors b/runs/baseline/checkpoint-3240/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9da137a2f1d4f3044d91a20bc418ad19587c3a50 --- /dev/null +++ b/runs/baseline/checkpoint-3240/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb208b9b92f9018d170f14ed359a3e6c33dc4579694afd0e56e77fb2daf1d8d3 +size 583356232 diff --git a/runs/baseline/checkpoint-3240/optimizer.pt b/runs/baseline/checkpoint-3240/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f5ae832414798980d79fc475e26a5035045e7b0e --- /dev/null +++ b/runs/baseline/checkpoint-3240/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:678f6eb482541f0d4bd629fe6b67a25e2edb11e98bee028d66fb82583a5c94d2 +size 1166825803 diff --git a/runs/baseline/checkpoint-3240/rng_state_0.pth b/runs/baseline/checkpoint-3240/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..854d237a4ff7656e8561049ef131e53b73d26e13 --- /dev/null +++ b/runs/baseline/checkpoint-3240/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be1f0fdefddf6f30886458ed628f8251f23fec39a294e6649a7ab241784b8460 +size 14917 diff --git a/runs/baseline/checkpoint-3240/rng_state_1.pth b/runs/baseline/checkpoint-3240/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..bf9f97063d7490c565f537598baa2da46c4b6b26 --- /dev/null +++ b/runs/baseline/checkpoint-3240/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35669f76342834bd66ba665f73a6cfdbe07c4f2c0f41f6e5e6ed2ad3ae67a161 +size 14917 diff --git a/runs/baseline/checkpoint-3240/scheduler.pt b/runs/baseline/checkpoint-3240/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1b3635513c9f21d830135bfbbf11f168f3565522 --- /dev/null +++ b/runs/baseline/checkpoint-3240/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:778f99daa1d0c600722ad8d85ac0f8172450ce32206d332c7edef1f0121201c6 +size 1465 diff --git a/runs/baseline/checkpoint-3240/tokenizer.json b/runs/baseline/checkpoint-3240/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-3240/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-3240/tokenizer_config.json b/runs/baseline/checkpoint-3240/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-3240/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-3240/trainer_state.json b/runs/baseline/checkpoint-3240/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..b338983818642f2c7c46b66ef54400bf6c2c9f96 --- /dev/null +++ b/runs/baseline/checkpoint-3240/trainer_state.json @@ -0,0 +1,23066 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 45.0, + "eval_steps": 500, + "global_step": 3240, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.641542434692383, + "eval_runtime": 51.6342, + "eval_samples_per_second": 47.294, + "eval_steps_per_second": 1.491, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.009694516658782959, + "learning_rate": 0.0020492498088767836, + "loss": 4.32808256149292, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.0097952950745821, + "learning_rate": 0.002048005909777114, + "loss": 4.280096054077148, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.010393359698355198, + "learning_rate": 0.0020467615812442106, + "loss": 4.317824363708496, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.00977372657507658, + "learning_rate": 0.0020455168242806054, + "loss": 4.3445329666137695, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.008568719029426575, + "learning_rate": 0.0020442716398891775, + "loss": 4.318026542663574, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.009732510894536972, + "learning_rate": 0.0020430260290731487, + "loss": 4.330609321594238, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.010429969057440758, + "learning_rate": 0.0020417799928360864, + "loss": 4.316794395446777, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.009292383678257465, + "learning_rate": 0.002040533532181898, + "loss": 4.317420959472656, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.011222455650568008, + "learning_rate": 0.0020392866481148356, + "loss": 4.292610168457031, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.013591725379228592, + "learning_rate": 0.0020380393416394894, + "loss": 4.325689315795898, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.013572026044130325, + "learning_rate": 0.002036791613760794, + "loss": 4.334960460662842, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.012986521236598492, + "learning_rate": 0.0020355434654840197, + "loss": 4.33829402923584, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.013321424834430218, + "learning_rate": 0.0020342948978147777, + "loss": 4.2926926612854, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.012307215481996536, + "learning_rate": 0.0020330459117590153, + "loss": 4.32757568359375, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.012909360229969025, + "learning_rate": 0.0020317965083230197, + "loss": 4.362889289855957, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017746610566973686, + "learning_rate": 0.0020305466885134115, + "loss": 4.3426313400268555, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.01696564070880413, + "learning_rate": 0.002029296453337148, + "loss": 4.3298516273498535, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.013348255306482315, + "learning_rate": 0.002028045803801522, + "loss": 4.360220432281494, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.013225145637989044, + "learning_rate": 0.002026794740914158, + "loss": 4.346250057220459, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.01569020375609398, + "learning_rate": 0.002025543265683016, + "loss": 4.308025360107422, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.018612321466207504, + "learning_rate": 0.002024291379116385, + "loss": 4.330190181732178, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.01619943603873253, + "learning_rate": 0.00202303908222289, + "loss": 4.339552402496338, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.015385213308036327, + "learning_rate": 0.0020217863760114826, + "loss": 4.3412184715271, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.018517112359404564, + "learning_rate": 0.002020533261491446, + "loss": 4.3289313316345215, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.014868509024381638, + "learning_rate": 0.0020192797396723903, + "loss": 4.359846115112305, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.014240248128771782, + "learning_rate": 0.0020180258115642578, + "loss": 4.369707107543945, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.018113158643245697, + "learning_rate": 0.0020167714781773143, + "loss": 4.362668514251709, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.01802460104227066, + "learning_rate": 0.002015516740522154, + "loss": 4.344742298126221, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.01414549071341753, + "learning_rate": 0.0020142615996096964, + "loss": 4.358080863952637, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.012157700955867767, + "learning_rate": 0.002013006056451186, + "loss": 4.347763538360596, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.011268534697592258, + "learning_rate": 0.0020117501120581905, + "loss": 4.335859298706055, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.010395628400146961, + "learning_rate": 0.002010493767442602, + "loss": 4.341609001159668, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.012147950939834118, + "learning_rate": 0.002009237023616635, + "loss": 4.328502655029297, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.00945813488215208, + "learning_rate": 0.0020079798815928252, + "loss": 4.311186790466309, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.009489500895142555, + "learning_rate": 0.0020067223423840294, + "loss": 4.371470928192139, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.009360029362142086, + "learning_rate": 0.0020054644070034234, + "loss": 4.33941125869751, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.010032779537141323, + "learning_rate": 0.0020042060764645037, + "loss": 4.365797519683838, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.009367046877741814, + "learning_rate": 0.002002947351781084, + "loss": 4.29443359375, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.008369733579456806, + "learning_rate": 0.0020016882339672967, + "loss": 4.294233322143555, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.010394785553216934, + "learning_rate": 0.0020004287240375893, + "loss": 4.352355003356934, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01068816613405943, + "learning_rate": 0.001999168823006728, + "loss": 4.342516899108887, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.009975973516702652, + "learning_rate": 0.0019979085318897905, + "loss": 4.346763610839844, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.009618151932954788, + "learning_rate": 0.001996647851702171, + "loss": 4.329115867614746, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.008399694226682186, + "learning_rate": 0.001995386783459577, + "loss": 4.28736686706543, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.008235268294811249, + "learning_rate": 0.001994125328178029, + "loss": 4.339872360229492, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.00836547277867794, + "learning_rate": 0.0019928634868738577, + "loss": 4.328156471252441, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.009619626216590405, + "learning_rate": 0.001991601260563706, + "loss": 4.337394714355469, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.009814038872718811, + "learning_rate": 0.0019903386502645275, + "loss": 4.325926780700684, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.009588954970240593, + "learning_rate": 0.001989075656993584, + "loss": 4.286477088928223, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.010471943765878677, + "learning_rate": 0.001987812281768447, + "loss": 4.322466850280762, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.010239035822451115, + "learning_rate": 0.001986548525606994, + "loss": 4.3487653732299805, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.008935568854212761, + "learning_rate": 0.001985284389527411, + "loss": 4.303601264953613, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.008624759502708912, + "learning_rate": 0.00198401987454819, + "loss": 4.34036922454834, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.009296547621488571, + "learning_rate": 0.0019827549816881277, + "loss": 4.273357391357422, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.009071138687431812, + "learning_rate": 0.0019814897119663248, + "loss": 4.321051597595215, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.008341389708220959, + "learning_rate": 0.0019802240664021876, + "loss": 4.343242645263672, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.007881715893745422, + "learning_rate": 0.0019789580460154227, + "loss": 4.32067346572876, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.008750553242862225, + "learning_rate": 0.0019776916518260404, + "loss": 4.330263614654541, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.008591460064053535, + "learning_rate": 0.001976424884854351, + "loss": 4.311957359313965, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.008084543980658054, + "learning_rate": 0.001975157746120966, + "loss": 4.331319808959961, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.008036594837903976, + "learning_rate": 0.001973890236646797, + "loss": 4.382656097412109, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.00805772002786398, + "learning_rate": 0.0019726223574530527, + "loss": 4.297247886657715, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.009019420482218266, + "learning_rate": 0.0019713541095612405, + "loss": 4.326526165008545, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.009953883476555347, + "learning_rate": 0.0019700854939931653, + "loss": 4.327356338500977, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.009340492077171803, + "learning_rate": 0.0019688165117709273, + "loss": 4.343993186950684, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.010395721532404423, + "learning_rate": 0.0019675471639169226, + "loss": 4.284934997558594, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.009946230798959732, + "learning_rate": 0.0019662774514538415, + "loss": 4.288936614990234, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.010981100611388683, + "learning_rate": 0.0019650073754046695, + "loss": 4.339935779571533, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.009688783437013626, + "learning_rate": 0.0019637369367926827, + "loss": 4.381425857543945, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.009442255832254887, + "learning_rate": 0.001962466136641452, + "loss": 4.304185390472412, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.009371291846036911, + "learning_rate": 0.0019611949759748364, + "loss": 4.306120872497559, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.009544814005494118, + "learning_rate": 0.001959923455816988, + "loss": 4.326251983642578, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.602745532989502, + "eval_runtime": 53.5041, + "eval_samples_per_second": 45.641, + "eval_steps_per_second": 1.439, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.01010364294052124, + "learning_rate": 0.0019586515771923474, + "loss": 4.24491024017334, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.012722729705274105, + "learning_rate": 0.0019573793411256446, + "loss": 4.257739543914795, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.01307204831391573, + "learning_rate": 0.0019561067486418967, + "loss": 4.291684150695801, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.011215966194868088, + "learning_rate": 0.0019548338007664098, + "loss": 4.237947463989258, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.012943019159138203, + "learning_rate": 0.0019535604985247727, + "loss": 4.242859363555908, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.01337282732129097, + "learning_rate": 0.0019522868429428636, + "loss": 4.238934516906738, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.013697991147637367, + "learning_rate": 0.0019510128350468438, + "loss": 4.285760402679443, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.012182159349322319, + "learning_rate": 0.001949738475863158, + "loss": 4.290044784545898, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.012620339170098305, + "learning_rate": 0.0019484637664185346, + "loss": 4.267187118530273, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.011284958571195602, + "learning_rate": 0.0019471887077399837, + "loss": 4.26662540435791, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.01161450706422329, + "learning_rate": 0.001945913300854797, + "loss": 4.283192157745361, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.010022968985140324, + "learning_rate": 0.001944637546790548, + "loss": 4.295843124389648, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.010341878980398178, + "learning_rate": 0.001943361446575087, + "loss": 4.295636177062988, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.011549185030162334, + "learning_rate": 0.0019420850012365457, + "loss": 4.288768768310547, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.012518810108304024, + "learning_rate": 0.0019408082118033331, + "loss": 4.229338645935059, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.012035328894853592, + "learning_rate": 0.0019395310793041353, + "loss": 4.258639335632324, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.011193782091140747, + "learning_rate": 0.0019382536047679146, + "loss": 4.239264488220215, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.009930136613547802, + "learning_rate": 0.0019369757892239096, + "loss": 4.252286434173584, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.00953229982405901, + "learning_rate": 0.0019356976337016327, + "loss": 4.293435096740723, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.010618340224027634, + "learning_rate": 0.0019344191392308716, + "loss": 4.235588550567627, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.008806651458144188, + "learning_rate": 0.0019331403068416856, + "loss": 4.26716423034668, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.008934796787798405, + "learning_rate": 0.0019318611375644065, + "loss": 4.258462905883789, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.009283283725380898, + "learning_rate": 0.0019305816324296386, + "loss": 4.257868766784668, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.009224812500178814, + "learning_rate": 0.0019293017924682555, + "loss": 4.258486747741699, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.008957969956099987, + "learning_rate": 0.0019280216187114016, + "loss": 4.309563636779785, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.010284051299095154, + "learning_rate": 0.0019267411121904894, + "loss": 4.275271415710449, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.009167762473225594, + "learning_rate": 0.0019254602739372001, + "loss": 4.238452911376953, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.009411636739969254, + "learning_rate": 0.0019241791049834816, + "loss": 4.2735595703125, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.009044171310961246, + "learning_rate": 0.0019228976063615497, + "loss": 4.27017879486084, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.008840068243443966, + "learning_rate": 0.0019216157791038834, + "loss": 4.28773307800293, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.009512359276413918, + "learning_rate": 0.001920333624243228, + "loss": 4.266008377075195, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.009946667589247227, + "learning_rate": 0.001919051142812593, + "loss": 4.279062271118164, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.00899894442409277, + "learning_rate": 0.0019177683358452505, + "loss": 4.277679443359375, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.00909668579697609, + "learning_rate": 0.0019164852043747338, + "loss": 4.311397552490234, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.009787327609956264, + "learning_rate": 0.0019152017494348408, + "loss": 4.30875825881958, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.008983899839222431, + "learning_rate": 0.001913917972059626, + "loss": 4.259649276733398, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.008411582559347153, + "learning_rate": 0.0019126338732834071, + "loss": 4.259811878204346, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.009162918664515018, + "learning_rate": 0.0019113494541407579, + "loss": 4.271515846252441, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.011304561980068684, + "learning_rate": 0.0019100647156665125, + "loss": 4.269217491149902, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.012483394704759121, + "learning_rate": 0.0019087796588957617, + "loss": 4.2949299812316895, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.012091380544006824, + "learning_rate": 0.0019074942848638522, + "loss": 4.289752006530762, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.010011567734181881, + "learning_rate": 0.0019062085946063869, + "loss": 4.294095993041992, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.008766217157244682, + "learning_rate": 0.0019049225891592226, + "loss": 4.256425857543945, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.012862488627433777, + "learning_rate": 0.001903636269558471, + "loss": 4.287847995758057, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.013917063362896442, + "learning_rate": 0.0019023496368404972, + "loss": 4.282784938812256, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.012129023671150208, + "learning_rate": 0.0019010626920419173, + "loss": 4.2486419677734375, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.010348426178097725, + "learning_rate": 0.0018997754361995986, + "loss": 4.289836883544922, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.011821693740785122, + "learning_rate": 0.001898487870350662, + "loss": 4.214344024658203, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.013136724010109901, + "learning_rate": 0.0018971999955324742, + "loss": 4.24208402633667, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.013173168525099754, + "learning_rate": 0.0018959118127826538, + "loss": 4.270164489746094, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.011018159799277782, + "learning_rate": 0.0018946233231390656, + "loss": 4.2571916580200195, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.01018658559769392, + "learning_rate": 0.0018933345276398227, + "loss": 4.283267498016357, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.009412688203155994, + "learning_rate": 0.0018920454273232848, + "loss": 4.272198677062988, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.007799267303198576, + "learning_rate": 0.0018907560232280557, + "loss": 4.260125637054443, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.008747464045882225, + "learning_rate": 0.0018894663163929863, + "loss": 4.2539215087890625, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.008003434166312218, + "learning_rate": 0.0018881763078571689, + "loss": 4.245024681091309, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.007593846879899502, + "learning_rate": 0.0018868859986599415, + "loss": 4.237929344177246, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.008099420927464962, + "learning_rate": 0.0018855953898408815, + "loss": 4.2789106369018555, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.008270486257970333, + "learning_rate": 0.0018843044824398097, + "loss": 4.263522148132324, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.009260379709303379, + "learning_rate": 0.0018830132774967866, + "loss": 4.275835037231445, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.009289492852985859, + "learning_rate": 0.0018817217760521132, + "loss": 4.313778400421143, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.009493921883404255, + "learning_rate": 0.0018804299791463281, + "loss": 4.303672790527344, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.009020023047924042, + "learning_rate": 0.0018791378878202105, + "loss": 4.245473384857178, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.009446556679904461, + "learning_rate": 0.0018778455031147724, + "loss": 4.267327785491943, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.009499974548816681, + "learning_rate": 0.0018765528260712672, + "loss": 4.301205635070801, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.008399393409490585, + "learning_rate": 0.00187525985773118, + "loss": 4.307802677154541, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.009519701823592186, + "learning_rate": 0.0018739665991362327, + "loss": 4.284796714782715, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01125232595950365, + "learning_rate": 0.0018726730513283804, + "loss": 4.285930633544922, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.011608537286520004, + "learning_rate": 0.00187137921534981, + "loss": 4.257689476013184, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.01064533181488514, + "learning_rate": 0.001870085092242944, + "loss": 4.249063968658447, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.009854599833488464, + "learning_rate": 0.0018687906830504321, + "loss": 4.255159378051758, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.01121739111840725, + "learning_rate": 0.0018674959888151576, + "loss": 4.287777900695801, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.5675368309021, + "eval_runtime": 51.8555, + "eval_samples_per_second": 47.092, + "eval_steps_per_second": 1.485, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.0104464590549469, + "learning_rate": 0.0018662010105802314, + "loss": 4.206080436706543, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.012800181284546852, + "learning_rate": 0.0018649057493889945, + "loss": 4.213550567626953, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.013382901437580585, + "learning_rate": 0.001863610206285015, + "loss": 4.190858840942383, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.014347050338983536, + "learning_rate": 0.0018623143823120894, + "loss": 4.22884464263916, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.01347152329981327, + "learning_rate": 0.001861018278514239, + "loss": 4.215579032897949, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.012786377221345901, + "learning_rate": 0.0018597218959357112, + "loss": 4.174999237060547, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.014137202873826027, + "learning_rate": 0.001858425235620979, + "loss": 4.21709680557251, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.013864399865269661, + "learning_rate": 0.001857128298614737, + "loss": 4.2375030517578125, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.010397658683359623, + "learning_rate": 0.0018558310859619048, + "loss": 4.231938362121582, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.00888819433748722, + "learning_rate": 0.0018545335987076223, + "loss": 4.195562839508057, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.010698067955672741, + "learning_rate": 0.001853235837897252, + "loss": 4.22847318649292, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.00987589918076992, + "learning_rate": 0.0018519378045763766, + "loss": 4.2147040367126465, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.011451782658696175, + "learning_rate": 0.001850639499790798, + "loss": 4.179263114929199, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.011224704794585705, + "learning_rate": 0.0018493409245865368, + "loss": 4.23030948638916, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.00996861606836319, + "learning_rate": 0.0018480420800098313, + "loss": 4.194283962249756, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.012143119238317013, + "learning_rate": 0.0018467429671071374, + "loss": 4.221704959869385, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.013126783072948456, + "learning_rate": 0.0018454435869251268, + "loss": 4.187748908996582, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.013031953014433384, + "learning_rate": 0.001844143940510687, + "loss": 4.248250961303711, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.012868993915617466, + "learning_rate": 0.001842844028910919, + "loss": 4.20161247253418, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.010102041065692902, + "learning_rate": 0.0018415438531731395, + "loss": 4.213654518127441, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.009599820710718632, + "learning_rate": 0.0018402434143448747, + "loss": 4.206313133239746, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.011536866426467896, + "learning_rate": 0.0018389427134738658, + "loss": 4.237072944641113, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.009558089077472687, + "learning_rate": 0.0018376417516080636, + "loss": 4.182195663452148, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01056219544261694, + "learning_rate": 0.0018363405297956302, + "loss": 4.232585906982422, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.012673584744334221, + "learning_rate": 0.0018350390490849354, + "loss": 4.228830337524414, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.011215884238481522, + "learning_rate": 0.0018337373105245598, + "loss": 4.216056823730469, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.011281915940344334, + "learning_rate": 0.00183243531516329, + "loss": 4.233404159545898, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.010381227359175682, + "learning_rate": 0.0018311330640501204, + "loss": 4.182013034820557, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.00900198519229889, + "learning_rate": 0.0018298305582342513, + "loss": 4.215488433837891, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.010198601521551609, + "learning_rate": 0.0018285277987650876, + "loss": 4.19389533996582, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.009965402074158192, + "learning_rate": 0.0018272247866922401, + "loss": 4.218005657196045, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.009299935773015022, + "learning_rate": 0.0018259215230655202, + "loss": 4.223684310913086, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.008977951481938362, + "learning_rate": 0.0018246180089349462, + "loss": 4.215778350830078, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.009261307306587696, + "learning_rate": 0.0018233142453507344, + "loss": 4.215066909790039, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.008322088047862053, + "learning_rate": 0.0018220102333633044, + "loss": 4.2188005447387695, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.008716144599020481, + "learning_rate": 0.0018207059740232739, + "loss": 4.187867641448975, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.008897020481526852, + "learning_rate": 0.0018194014683814625, + "loss": 4.218034267425537, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.007298174314200878, + "learning_rate": 0.0018180967174888857, + "loss": 4.2017011642456055, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.008491789922118187, + "learning_rate": 0.0018167917223967582, + "loss": 4.239873886108398, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.008413378149271011, + "learning_rate": 0.0018154864841564908, + "loss": 4.234771251678467, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.008141471073031425, + "learning_rate": 0.0018141810038196909, + "loss": 4.185944080352783, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.009480944834649563, + "learning_rate": 0.0018128752824381603, + "loss": 4.206586837768555, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.0106047997251153, + "learning_rate": 0.001811569321063894, + "loss": 4.235287666320801, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.010663514956831932, + "learning_rate": 0.0018102631207490838, + "loss": 4.204899311065674, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.011473523452877998, + "learning_rate": 0.001808956682546109, + "loss": 4.253668308258057, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01165992021560669, + "learning_rate": 0.0018076500075075456, + "loss": 4.209845542907715, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.011692410334944725, + "learning_rate": 0.0018063430966861564, + "loss": 4.250368595123291, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01061506848782301, + "learning_rate": 0.0018050359511348965, + "loss": 4.193821907043457, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.010510941036045551, + "learning_rate": 0.0018037285719069097, + "loss": 4.254327774047852, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.011587041430175304, + "learning_rate": 0.001802420960055528, + "loss": 4.2838945388793945, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.01257420051842928, + "learning_rate": 0.0018011131166342693, + "loss": 4.258241653442383, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.00936233438551426, + "learning_rate": 0.0017998050426968407, + "loss": 4.201854705810547, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.010093127377331257, + "learning_rate": 0.0017984967392971338, + "loss": 4.193051338195801, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.010032009333372116, + "learning_rate": 0.0017971882074892236, + "loss": 4.230212211608887, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.009455587714910507, + "learning_rate": 0.0017958794483273726, + "loss": 4.2157673835754395, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.009549238719046116, + "learning_rate": 0.001794570462866023, + "loss": 4.260272026062012, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.011059699580073357, + "learning_rate": 0.0017932612521598013, + "loss": 4.214818000793457, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.011653627268970013, + "learning_rate": 0.0017919518172635146, + "loss": 4.205112457275391, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.012438314035534859, + "learning_rate": 0.001790642159232151, + "loss": 4.179045677185059, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.011234673671424389, + "learning_rate": 0.0017893322791208784, + "loss": 4.239317417144775, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.010923845693469048, + "learning_rate": 0.0017880221779850434, + "loss": 4.230571269989014, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.009219309315085411, + "learning_rate": 0.001786711856880171, + "loss": 4.229068756103516, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.009849398396909237, + "learning_rate": 0.0017854013168619638, + "loss": 4.254772186279297, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.01117763202637434, + "learning_rate": 0.001784090558986299, + "loss": 4.252647399902344, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.010664612054824829, + "learning_rate": 0.0017827795843092316, + "loss": 4.204596519470215, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.00847950391471386, + "learning_rate": 0.0017814683938869895, + "loss": 4.211061954498291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.007305726408958435, + "learning_rate": 0.0017801569887759753, + "loss": 4.181221008300781, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.007970666512846947, + "learning_rate": 0.0017788453700327653, + "loss": 4.237855434417725, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.008873614482581615, + "learning_rate": 0.0017775335387141054, + "loss": 4.229188919067383, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.009691286832094193, + "learning_rate": 0.0017762214958769156, + "loss": 4.205491542816162, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.009360672906041145, + "learning_rate": 0.0017749092425782852, + "loss": 4.233466625213623, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.009895884431898594, + "learning_rate": 0.0017735967798754723, + "loss": 4.187763214111328, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.535672187805176, + "eval_runtime": 53.1736, + "eval_samples_per_second": 45.925, + "eval_steps_per_second": 1.448, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.00966129545122385, + "learning_rate": 0.0017722841088259045, + "loss": 4.1451735496521, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.010733380913734436, + "learning_rate": 0.0017709712304871774, + "loss": 4.1616363525390625, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.0117328567430377, + "learning_rate": 0.0017696581459170535, + "loss": 4.126331806182861, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.012640298344194889, + "learning_rate": 0.0017683448561734607, + "loss": 4.163730621337891, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.013074076734483242, + "learning_rate": 0.0017670313623144934, + "loss": 4.199502944946289, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.01132719125598669, + "learning_rate": 0.0017657176653984097, + "loss": 4.143380165100098, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.010008333250880241, + "learning_rate": 0.0017644037664836319, + "loss": 4.170320510864258, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.011090578511357307, + "learning_rate": 0.0017630896666287438, + "loss": 4.1571478843688965, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.011047456413507462, + "learning_rate": 0.0017617753668924924, + "loss": 4.101892948150635, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.010098653845489025, + "learning_rate": 0.0017604608683337857, + "loss": 4.154077529907227, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.010499860160052776, + "learning_rate": 0.00175914617201169, + "loss": 4.116868019104004, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.010939318686723709, + "learning_rate": 0.0017578312789854338, + "loss": 4.159144878387451, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.010622999630868435, + "learning_rate": 0.0017565161903144023, + "loss": 4.171411991119385, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.0108700692653656, + "learning_rate": 0.0017552009070581386, + "loss": 4.100748062133789, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.009794988669455051, + "learning_rate": 0.0017538854302763423, + "loss": 4.185161590576172, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.007750354241579771, + "learning_rate": 0.0017525697610288702, + "loss": 4.149083137512207, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.009923134930431843, + "learning_rate": 0.001751253900375732, + "loss": 4.205166816711426, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.010811141692101955, + "learning_rate": 0.0017499378493770936, + "loss": 4.131865978240967, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.010868688113987446, + "learning_rate": 0.0017486216090932733, + "loss": 4.148791790008545, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.011224126443266869, + "learning_rate": 0.001747305180584743, + "loss": 4.152503490447998, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.013262554071843624, + "learning_rate": 0.001745988564912124, + "loss": 4.164926528930664, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.013089599087834358, + "learning_rate": 0.001744671763136191, + "loss": 4.1626458168029785, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.01327646616846323, + "learning_rate": 0.0017433547763178665, + "loss": 4.189356803894043, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.014539042487740517, + "learning_rate": 0.0017420376055182233, + "loss": 4.176387786865234, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.013269098475575447, + "learning_rate": 0.001740720251798482, + "loss": 4.180716514587402, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.010922391898930073, + "learning_rate": 0.0017394027162200105, + "loss": 4.163928031921387, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.013355271890759468, + "learning_rate": 0.0017380849998443243, + "loss": 4.150012016296387, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.014149201102554798, + "learning_rate": 0.0017367671037330828, + "loss": 4.173713207244873, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01354879979044199, + "learning_rate": 0.0017354490289480918, + "loss": 4.150563716888428, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.013788704760372639, + "learning_rate": 0.0017341307765512998, + "loss": 4.213858604431152, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.011597222648561, + "learning_rate": 0.0017328123476047994, + "loss": 4.166989326477051, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.01055675558745861, + "learning_rate": 0.001731493743170824, + "loss": 4.172226428985596, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.011600397527217865, + "learning_rate": 0.0017301749643117506, + "loss": 4.1656951904296875, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.01114651933312416, + "learning_rate": 0.0017288560120900943, + "loss": 4.189825057983398, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.009811583906412125, + "learning_rate": 0.0017275368875685126, + "loss": 4.181164741516113, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.009758968837559223, + "learning_rate": 0.0017262175918097992, + "loss": 4.163188934326172, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.00999707542359829, + "learning_rate": 0.001724898125876887, + "loss": 4.184117317199707, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.009528180584311485, + "learning_rate": 0.001723578490832846, + "loss": 4.16407585144043, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.009717358276247978, + "learning_rate": 0.0017222586877408826, + "loss": 4.209892272949219, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.010909050703048706, + "learning_rate": 0.0017209387176643377, + "loss": 4.19977331161499, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.011234750039875507, + "learning_rate": 0.0017196185816666875, + "loss": 4.19363260269165, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.010300721041858196, + "learning_rate": 0.0017182982808115425, + "loss": 4.142427921295166, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.00941646471619606, + "learning_rate": 0.001716977816162644, + "loss": 4.159372329711914, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.009304054081439972, + "learning_rate": 0.0017156571887838679, + "loss": 4.207922458648682, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.009496180340647697, + "learning_rate": 0.0017143363997392184, + "loss": 4.137894630432129, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.009257436729967594, + "learning_rate": 0.0017130154500928326, + "loss": 4.198753833770752, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.00900586973875761, + "learning_rate": 0.0017116943409089746, + "loss": 4.160684585571289, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.009398494847118855, + "learning_rate": 0.00171037307325204, + "loss": 4.1819353103637695, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.008633013814687729, + "learning_rate": 0.0017090516481865483, + "loss": 4.176833152770996, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.00898425467312336, + "learning_rate": 0.0017077300667771484, + "loss": 4.190657138824463, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.008182398043572903, + "learning_rate": 0.0017064083300886155, + "loss": 4.173429489135742, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.009589143097400665, + "learning_rate": 0.0017050864391858478, + "loss": 4.1559014320373535, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.010304383002221584, + "learning_rate": 0.0017037643951338698, + "loss": 4.155453681945801, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.009943748824298382, + "learning_rate": 0.0017024421989978275, + "loss": 4.119734287261963, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.009238033555448055, + "learning_rate": 0.001701119851842992, + "loss": 4.188974857330322, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.008058690465986729, + "learning_rate": 0.0016997973547347531, + "loss": 4.19934606552124, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.008485953323543072, + "learning_rate": 0.0016984747087386233, + "loss": 4.165615558624268, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.008476605638861656, + "learning_rate": 0.0016971519149202347, + "loss": 4.21021842956543, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.008588708005845547, + "learning_rate": 0.0016958289743453385, + "loss": 4.208064556121826, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.009223178029060364, + "learning_rate": 0.0016945058880798037, + "loss": 4.158413887023926, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.009196167811751366, + "learning_rate": 0.0016931826571896166, + "loss": 4.203187465667725, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.009488921612501144, + "learning_rate": 0.001691859282740882, + "loss": 4.178227424621582, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.009253161028027534, + "learning_rate": 0.0016905357657998167, + "loss": 4.131230354309082, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.010069048963487148, + "learning_rate": 0.0016892121074327559, + "loss": 4.210734844207764, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.009945095516741276, + "learning_rate": 0.001687888308706147, + "loss": 4.184042930603027, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.010483640246093273, + "learning_rate": 0.0016865643706865505, + "loss": 4.156919479370117, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.011560715734958649, + "learning_rate": 0.0016852402944406391, + "loss": 4.185520172119141, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.010001277551054955, + "learning_rate": 0.0016839160810351978, + "loss": 4.226828575134277, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.010335934348404408, + "learning_rate": 0.0016825917315371202, + "loss": 4.183531284332275, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.011322776786983013, + "learning_rate": 0.001681267247013412, + "loss": 4.179635047912598, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.009663194417953491, + "learning_rate": 0.001679942628531186, + "loss": 4.165654182434082, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.010590246878564358, + "learning_rate": 0.0016786178771576643, + "loss": 4.1539459228515625, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.511386394500732, + "eval_runtime": 51.7751, + "eval_samples_per_second": 47.166, + "eval_steps_per_second": 1.487, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.01254354789853096, + "learning_rate": 0.0016772929939601728, + "loss": 4.104002475738525, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.01300987508147955, + "learning_rate": 0.0016759679800061483, + "loss": 4.127395153045654, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.013325816951692104, + "learning_rate": 0.0016746428363631299, + "loss": 4.131229400634766, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.014302870258688927, + "learning_rate": 0.001673317564098761, + "loss": 4.079358100891113, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.013776068575680256, + "learning_rate": 0.0016719921642807905, + "loss": 4.132065773010254, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.010788560844957829, + "learning_rate": 0.0016706666379770689, + "loss": 4.134626388549805, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.009766630828380585, + "learning_rate": 0.0016693409862555493, + "loss": 4.1645426750183105, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.011760029010474682, + "learning_rate": 0.0016680152101842842, + "loss": 4.110664367675781, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.011047836393117905, + "learning_rate": 0.0016666893108314288, + "loss": 4.129222869873047, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.009989139623939991, + "learning_rate": 0.0016653632892652344, + "loss": 4.076835632324219, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.010033783502876759, + "learning_rate": 0.001664037146554054, + "loss": 4.11555290222168, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.00996427796781063, + "learning_rate": 0.001662710883766337, + "loss": 4.131463527679443, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.01044023409485817, + "learning_rate": 0.0016613845019706288, + "loss": 4.098750114440918, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.010033872909843922, + "learning_rate": 0.0016600580022355712, + "loss": 4.161447048187256, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.009329567663371563, + "learning_rate": 0.0016587313856299012, + "loss": 4.146491527557373, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.010061589069664478, + "learning_rate": 0.0016574046532224504, + "loss": 4.107517719268799, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.009491910226643085, + "learning_rate": 0.0016560778060821424, + "loss": 4.098226547241211, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.008529345504939556, + "learning_rate": 0.0016547508452779943, + "loss": 4.1247711181640625, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.00936342403292656, + "learning_rate": 0.0016534237718791136, + "loss": 4.120573043823242, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.010255641303956509, + "learning_rate": 0.0016520965869547008, + "loss": 4.089067459106445, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.011638758704066277, + "learning_rate": 0.001650769291574044, + "loss": 4.074847221374512, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.014707857742905617, + "learning_rate": 0.0016494418868065214, + "loss": 4.153377532958984, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.01650111749768257, + "learning_rate": 0.0016481143737215986, + "loss": 4.134062767028809, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.014833372086286545, + "learning_rate": 0.00164678675338883, + "loss": 4.1344828605651855, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.012257886119186878, + "learning_rate": 0.0016454590268778538, + "loss": 4.098813533782959, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.013508322648704052, + "learning_rate": 0.0016441311952583959, + "loss": 4.106297492980957, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.01256900280714035, + "learning_rate": 0.001642803259600267, + "loss": 4.143631935119629, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.009885406121611595, + "learning_rate": 0.00164147522097336, + "loss": 4.120885848999023, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.012004674412310123, + "learning_rate": 0.0016401470804476518, + "loss": 4.089243412017822, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01224028691649437, + "learning_rate": 0.0016388188390932012, + "loss": 4.077765464782715, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.012511290609836578, + "learning_rate": 0.001637490497980149, + "loss": 4.161346435546875, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.010701716877520084, + "learning_rate": 0.0016361620581787143, + "loss": 4.10698938369751, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.010713113471865654, + "learning_rate": 0.001634833520759198, + "loss": 4.114283561706543, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.011018907651305199, + "learning_rate": 0.001633504886791978, + "loss": 4.134284019470215, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.011279584839940071, + "learning_rate": 0.0016321761573475117, + "loss": 4.1456804275512695, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.010642931796610355, + "learning_rate": 0.001630847333496332, + "loss": 4.160918235778809, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.011069678701460361, + "learning_rate": 0.0016295184163090482, + "loss": 4.136838912963867, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.011007166467607021, + "learning_rate": 0.0016281894068563443, + "loss": 4.129153728485107, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.010055703110992908, + "learning_rate": 0.0016268603062089802, + "loss": 4.138665676116943, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.009839230217039585, + "learning_rate": 0.0016255311154377875, + "loss": 4.1366496086120605, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.011191432364284992, + "learning_rate": 0.0016242018356136714, + "loss": 4.129880905151367, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.012130341492593288, + "learning_rate": 0.0016228724678076095, + "loss": 4.1425981521606445, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.012707795947790146, + "learning_rate": 0.0016215430130906482, + "loss": 4.115452766418457, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.012707809917628765, + "learning_rate": 0.001620213472533906, + "loss": 4.122758865356445, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.01182397548109293, + "learning_rate": 0.001618883847208569, + "loss": 4.140275955200195, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.010865340940654278, + "learning_rate": 0.001617554138185893, + "loss": 4.137759208679199, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.010973592288792133, + "learning_rate": 0.0016162243465371998, + "loss": 4.138947010040283, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.011247226968407631, + "learning_rate": 0.001614894473333879, + "loss": 4.115516662597656, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.010837487876415253, + "learning_rate": 0.0016135645196473853, + "loss": 4.160427093505859, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.009096396155655384, + "learning_rate": 0.0016122344865492385, + "loss": 4.1200032234191895, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.010671247728168964, + "learning_rate": 0.0016109043751110223, + "loss": 4.097919464111328, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.0107562355697155, + "learning_rate": 0.0016095741864043831, + "loss": 4.096290588378906, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.008909327909350395, + "learning_rate": 0.0016082439215010305, + "loss": 4.171017646789551, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.010305124334990978, + "learning_rate": 0.0016069135814727343, + "loss": 4.1450276374816895, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.011235613375902176, + "learning_rate": 0.0016055831673913255, + "loss": 4.134704113006592, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.00825632456690073, + "learning_rate": 0.0016042526803286949, + "loss": 4.107443332672119, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.009056378155946732, + "learning_rate": 0.0016029221213567921, + "loss": 4.132656097412109, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.008835164830088615, + "learning_rate": 0.001601591491547624, + "loss": 4.124729633331299, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.00915333442389965, + "learning_rate": 0.0016002607919732557, + "loss": 4.119500160217285, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.008532539941370487, + "learning_rate": 0.0015989300237058066, + "loss": 4.141661643981934, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.009952311404049397, + "learning_rate": 0.0015975991878174533, + "loss": 4.140929222106934, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.008920297026634216, + "learning_rate": 0.0015962682853804268, + "loss": 4.155877113342285, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.008770049549639225, + "learning_rate": 0.0015949373174670104, + "loss": 4.126394271850586, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.0078084105625748634, + "learning_rate": 0.0015936062851495411, + "loss": 4.128876209259033, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.00816641841083765, + "learning_rate": 0.0015922751895004078, + "loss": 4.137410640716553, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.008895459584891796, + "learning_rate": 0.0015909440315920505, + "loss": 4.07957649230957, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.008551528677344322, + "learning_rate": 0.001589612812496958, + "loss": 4.104213714599609, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.008325488306581974, + "learning_rate": 0.0015882815332876707, + "loss": 4.1584792137146, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.009341313503682613, + "learning_rate": 0.0015869501950367744, + "loss": 4.1417436599731445, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.00999050959944725, + "learning_rate": 0.0015856187988169067, + "loss": 4.136049270629883, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.008889932185411453, + "learning_rate": 0.0015842873457007484, + "loss": 4.16257381439209, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.009109573438763618, + "learning_rate": 0.0015829558367610268, + "loss": 4.171441555023193, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.485942840576172, + "eval_runtime": 52.1051, + "eval_samples_per_second": 46.867, + "eval_steps_per_second": 1.478, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.009881424717605114, + "learning_rate": 0.001581624273070515, + "loss": 4.059608459472656, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.011156721040606499, + "learning_rate": 0.0015802926557020294, + "loss": 4.047014236450195, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.012326992116868496, + "learning_rate": 0.0015789609857284309, + "loss": 4.06541109085083, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.01332208700478077, + "learning_rate": 0.0015776292642226213, + "loss": 4.042714595794678, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.012720128521323204, + "learning_rate": 0.0015762974922575446, + "loss": 4.098389625549316, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.010723215527832508, + "learning_rate": 0.0015749656709061854, + "loss": 4.059185981750488, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.010605281218886375, + "learning_rate": 0.001573633801241569, + "loss": 4.099538326263428, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.01101530622690916, + "learning_rate": 0.0015723018843367576, + "loss": 4.095674514770508, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.01135838869959116, + "learning_rate": 0.0015709699212648532, + "loss": 4.070535659790039, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.012119271792471409, + "learning_rate": 0.0015696379130989936, + "loss": 4.060605049133301, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.012783961370587349, + "learning_rate": 0.001568305860912355, + "loss": 4.075931072235107, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.011486220173537731, + "learning_rate": 0.0015669737657781461, + "loss": 4.075855255126953, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.01037552859634161, + "learning_rate": 0.0015656416287696136, + "loss": 4.087729454040527, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.009802743792533875, + "learning_rate": 0.0015643094509600354, + "loss": 4.0855607986450195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.010804438032209873, + "learning_rate": 0.001562977233422723, + "loss": 4.064528465270996, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.009505728259682655, + "learning_rate": 0.0015616449772310206, + "loss": 4.036775588989258, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.010634485632181168, + "learning_rate": 0.0015603126834583024, + "loss": 4.069923400878906, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.011392096057534218, + "learning_rate": 0.0015589803531779736, + "loss": 4.0812177658081055, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.011851722374558449, + "learning_rate": 0.0015576479874634683, + "loss": 4.09183406829834, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.011205444112420082, + "learning_rate": 0.0015563155873882502, + "loss": 4.07825231552124, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.009676349349319935, + "learning_rate": 0.0015549831540258093, + "loss": 4.052826404571533, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.009376165457069874, + "learning_rate": 0.0015536506884496638, + "loss": 4.02100133895874, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.00936955213546753, + "learning_rate": 0.0015523181917333564, + "loss": 4.049526214599609, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.00937784556299448, + "learning_rate": 0.0015509856649504564, + "loss": 4.096356391906738, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.010014008730649948, + "learning_rate": 0.0015496531091745555, + "loss": 4.03254508972168, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.009909634478390217, + "learning_rate": 0.0015483205254792705, + "loss": 4.067630767822266, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.009319501928985119, + "learning_rate": 0.0015469879149382399, + "loss": 4.094167709350586, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.008634721860289574, + "learning_rate": 0.0015456552786251244, + "loss": 4.122404098510742, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.009412454441189766, + "learning_rate": 0.0015443226176136049, + "loss": 4.081546783447266, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.010102108120918274, + "learning_rate": 0.0015429899329773814, + "loss": 4.049839019775391, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.010921654291450977, + "learning_rate": 0.0015416572257901746, + "loss": 4.097097396850586, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.010607713833451271, + "learning_rate": 0.0015403244971257226, + "loss": 4.084481239318848, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.011887960135936737, + "learning_rate": 0.0015389917480577806, + "loss": 4.074968338012695, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01089888159185648, + "learning_rate": 0.0015376589796601203, + "loss": 4.062132835388184, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.009934183210134506, + "learning_rate": 0.0015363261930065298, + "loss": 4.128092288970947, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.00878534372895956, + "learning_rate": 0.0015349933891708105, + "loss": 4.069051742553711, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.009928155690431595, + "learning_rate": 0.0015336605692267788, + "loss": 4.073912620544434, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.011091411113739014, + "learning_rate": 0.0015323277342482637, + "loss": 4.125944137573242, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.010314172133803368, + "learning_rate": 0.0015309948853091055, + "loss": 4.062620162963867, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.010423967614769936, + "learning_rate": 0.001529662023483158, + "loss": 4.101737976074219, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.009822997264564037, + "learning_rate": 0.0015283291498442817, + "loss": 4.064520359039307, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.009757095016539097, + "learning_rate": 0.001526996265466351, + "loss": 4.100624084472656, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.008650688454508781, + "learning_rate": 0.001525663371423246, + "loss": 4.109685897827148, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.007378234062343836, + "learning_rate": 0.0015243304687888553, + "loss": 4.044726848602295, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.008194336667656898, + "learning_rate": 0.001522997558637074, + "loss": 4.097923278808594, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.009764892049133778, + "learning_rate": 0.0015216646420418048, + "loss": 4.102148056030273, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.009242868050932884, + "learning_rate": 0.001520331720076953, + "loss": 4.068563461303711, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.010409324429929256, + "learning_rate": 0.001518998793816431, + "loss": 4.065944194793701, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.009215258061885834, + "learning_rate": 0.0015176658643341527, + "loss": 4.10223388671875, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.008065393194556236, + "learning_rate": 0.0015163329327040366, + "loss": 4.099432945251465, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.008498288691043854, + "learning_rate": 0.001515, + "loss": 4.109854698181152, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.008919150568544865, + "learning_rate": 0.0015136670672959633, + "loss": 4.128247261047363, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.009104429744184017, + "learning_rate": 0.0015123341356658474, + "loss": 4.068622589111328, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.008558365516364574, + "learning_rate": 0.0015110012061835692, + "loss": 4.114214897155762, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.007865076884627342, + "learning_rate": 0.0015096682799230477, + "loss": 4.11155366897583, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.007848534733057022, + "learning_rate": 0.0015083353579581958, + "loss": 4.045694351196289, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.008238030597567558, + "learning_rate": 0.0015070024413629264, + "loss": 4.093930721282959, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.007536398712545633, + "learning_rate": 0.0015056695312111448, + "loss": 4.106235504150391, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.008156892843544483, + "learning_rate": 0.001504336628576754, + "loss": 4.13060998916626, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.008144439198076725, + "learning_rate": 0.0015030037345336487, + "loss": 4.091304779052734, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.007757358253002167, + "learning_rate": 0.001501670850155718, + "loss": 4.069276809692383, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.00860623549669981, + "learning_rate": 0.0015003379765168425, + "loss": 4.107658386230469, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.008614457212388515, + "learning_rate": 0.0014990051146908949, + "loss": 4.106705665588379, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.009120592847466469, + "learning_rate": 0.001497672265751737, + "loss": 4.0891923904418945, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.009446980431675911, + "learning_rate": 0.0014963394307732213, + "loss": 4.098847389221191, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.010494766756892204, + "learning_rate": 0.0014950066108291894, + "loss": 4.115325450897217, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.010057658888399601, + "learning_rate": 0.0014936738069934706, + "loss": 4.099984169006348, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.009071771055459976, + "learning_rate": 0.0014923410203398796, + "loss": 4.091543674468994, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.009937075898051262, + "learning_rate": 0.00149100825194222, + "loss": 4.086492538452148, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.010222986340522766, + "learning_rate": 0.0014896755028742778, + "loss": 4.063841819763184, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.010513762012124062, + "learning_rate": 0.001488342774209826, + "loss": 4.11991548538208, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.009530222974717617, + "learning_rate": 0.0014870100670226192, + "loss": 4.043334007263184, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.467459678649902, + "eval_runtime": 54.1628, + "eval_samples_per_second": 45.086, + "eval_steps_per_second": 1.422, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.010274187661707401, + "learning_rate": 0.0014856773823863955, + "loss": 4.049561500549316, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.011448442004621029, + "learning_rate": 0.001484344721374876, + "loss": 4.019484996795654, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.011361192911863327, + "learning_rate": 0.0014830120850617599, + "loss": 4.0429887771606445, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.010250156745314598, + "learning_rate": 0.0014816794745207299, + "loss": 4.012682914733887, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.010024400427937508, + "learning_rate": 0.0014803468908254449, + "loss": 3.979409694671631, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.01153535582125187, + "learning_rate": 0.0014790143350495446, + "loss": 4.045780181884766, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.012632817961275578, + "learning_rate": 0.001477681808266644, + "loss": 4.057641983032227, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.013758293353021145, + "learning_rate": 0.0014763493115503366, + "loss": 4.061607360839844, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.014041583053767681, + "learning_rate": 0.001475016845974191, + "loss": 4.035029411315918, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.013629456982016563, + "learning_rate": 0.0014736844126117497, + "loss": 4.062169075012207, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.011023213155567646, + "learning_rate": 0.0014723520125365323, + "loss": 4.02731990814209, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.01001750584691763, + "learning_rate": 0.0014710196468220266, + "loss": 4.019882678985596, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.008321646600961685, + "learning_rate": 0.001469687316541698, + "loss": 4.005446434020996, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.008935946971178055, + "learning_rate": 0.0014683550227689796, + "loss": 4.017102241516113, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.009137130342423916, + "learning_rate": 0.001467022766577277, + "loss": 4.024102210998535, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.00805022194981575, + "learning_rate": 0.001465690549039965, + "loss": 3.9935250282287598, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.009204058907926083, + "learning_rate": 0.0014643583712303865, + "loss": 4.042818546295166, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.009812815114855766, + "learning_rate": 0.001463026234221854, + "loss": 4.0252485275268555, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.009955782443284988, + "learning_rate": 0.0014616941390876454, + "loss": 3.9839768409729004, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.009374645538628101, + "learning_rate": 0.001460362086901007, + "loss": 4.040245056152344, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.010008739307522774, + "learning_rate": 0.0014590300787351471, + "loss": 4.0306596755981445, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.009014200419187546, + "learning_rate": 0.001457698115663243, + "loss": 4.028713226318359, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.008381523191928864, + "learning_rate": 0.0014563661987584313, + "loss": 4.024718284606934, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.008627948351204395, + "learning_rate": 0.0014550343290938143, + "loss": 4.02303409576416, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.010257705114781857, + "learning_rate": 0.0014537025077424556, + "loss": 4.061058044433594, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.010759209282696247, + "learning_rate": 0.0014523707357773788, + "loss": 4.064718246459961, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.009966139681637287, + "learning_rate": 0.0014510390142715697, + "loss": 3.9919447898864746, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.00997584406286478, + "learning_rate": 0.001449707344297971, + "loss": 4.0431599617004395, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.010846944525837898, + "learning_rate": 0.0014483757269294858, + "loss": 4.0222697257995605, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.00991352554410696, + "learning_rate": 0.0014470441632389735, + "loss": 4.059438705444336, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.008994397707283497, + "learning_rate": 0.0014457126542992518, + "loss": 4.021494388580322, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.009567267261445522, + "learning_rate": 0.0014443812011830934, + "loss": 4.070810317993164, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.009199011139571667, + "learning_rate": 0.0014430498049632255, + "loss": 4.020938873291016, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.010440350510179996, + "learning_rate": 0.0014417184667123303, + "loss": 4.062529563903809, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.009962482377886772, + "learning_rate": 0.0014403871875030424, + "loss": 4.0687150955200195, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.009389379993081093, + "learning_rate": 0.0014390559684079502, + "loss": 4.055185794830322, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.011469941586256027, + "learning_rate": 0.0014377248104995924, + "loss": 4.041023254394531, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.011034650728106499, + "learning_rate": 0.0014363937148504588, + "loss": 4.054254531860352, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.010651151649653912, + "learning_rate": 0.0014350626825329897, + "loss": 4.037753582000732, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.010747930034995079, + "learning_rate": 0.0014337317146195733, + "loss": 4.059938907623291, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.00987576600164175, + "learning_rate": 0.0014324008121825468, + "loss": 4.075366973876953, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.008504371158778667, + "learning_rate": 0.0014310699762941936, + "loss": 4.026755332946777, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.009445052593946457, + "learning_rate": 0.001429739208026745, + "loss": 4.034431457519531, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.010560023598372936, + "learning_rate": 0.0014284085084523763, + "loss": 4.101158618927002, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.009920580312609673, + "learning_rate": 0.0014270778786432078, + "loss": 4.041321754455566, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.009854268282651901, + "learning_rate": 0.0014257473196713053, + "loss": 4.049882411956787, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.009825135581195354, + "learning_rate": 0.0014244168326086745, + "loss": 4.067440986633301, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.010506784543395042, + "learning_rate": 0.0014230864185272663, + "loss": 4.042476654052734, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.009527767077088356, + "learning_rate": 0.0014217560784989696, + "loss": 4.055248260498047, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.01013826485723257, + "learning_rate": 0.001420425813595617, + "loss": 4.05246639251709, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.011020111851394176, + "learning_rate": 0.0014190956248889778, + "loss": 4.032987117767334, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.010647117160260677, + "learning_rate": 0.0014177655134507614, + "loss": 4.093589782714844, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.009484780952334404, + "learning_rate": 0.001416435480352615, + "loss": 4.017437934875488, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.0114199910312891, + "learning_rate": 0.0014151055266661211, + "loss": 4.080643653869629, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.011137664318084717, + "learning_rate": 0.0014137756534628007, + "loss": 4.069380760192871, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.008972682990133762, + "learning_rate": 0.0014124458618141073, + "loss": 4.051383018493652, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.010311643593013287, + "learning_rate": 0.0014111161527914317, + "loss": 4.048953533172607, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.01113453321158886, + "learning_rate": 0.0014097865274660944, + "loss": 4.063544273376465, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.01091703586280346, + "learning_rate": 0.0014084569869093518, + "loss": 4.006721496582031, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.010672449134290218, + "learning_rate": 0.0014071275321923906, + "loss": 4.020220756530762, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.009682527743279934, + "learning_rate": 0.001405798164386328, + "loss": 4.057307243347168, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.009808676317334175, + "learning_rate": 0.0014044688845622126, + "loss": 4.105412483215332, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.010351189412176609, + "learning_rate": 0.0014031396937910202, + "loss": 4.066454887390137, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.010038275271654129, + "learning_rate": 0.001401810593143656, + "loss": 4.056752681732178, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01016094908118248, + "learning_rate": 0.001400481583690952, + "loss": 4.097933292388916, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.008683547377586365, + "learning_rate": 0.0013991526665036679, + "loss": 4.0386152267456055, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.009556332603096962, + "learning_rate": 0.0013978238426524882, + "loss": 4.05835485458374, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.009597044438123703, + "learning_rate": 0.001396495113208022, + "loss": 4.052363872528076, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.010021139867603779, + "learning_rate": 0.0013951664792408026, + "loss": 4.083431243896484, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.010611358098685741, + "learning_rate": 0.0013938379418212863, + "loss": 4.042330265045166, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.009132112376391888, + "learning_rate": 0.001392509502019852, + "loss": 4.058825969696045, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.009808709844946861, + "learning_rate": 0.0013911811609067992, + "loss": 4.093010902404785, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.4473161697387695, + "eval_runtime": 51.7738, + "eval_samples_per_second": 47.167, + "eval_steps_per_second": 1.487, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.011859052814543247, + "learning_rate": 0.0013898529195523488, + "loss": 3.9572651386260986, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.013210238888859749, + "learning_rate": 0.0013885247790266406, + "loss": 3.9812350273132324, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.012981324456632137, + "learning_rate": 0.001387196740399733, + "loss": 3.987725019454956, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.01424382347613573, + "learning_rate": 0.001385868804741604, + "loss": 3.9719889163970947, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.0154642965644598, + "learning_rate": 0.0013845409731221465, + "loss": 3.984248161315918, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.014779271557927132, + "learning_rate": 0.0013832132466111706, + "loss": 4.002912521362305, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.012921188026666641, + "learning_rate": 0.0013818856262784013, + "loss": 3.9588499069213867, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01165325753390789, + "learning_rate": 0.0013805581131934792, + "loss": 3.993288993835449, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.008908954448997974, + "learning_rate": 0.0013792307084259563, + "loss": 3.987151861190796, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.012040507979691029, + "learning_rate": 0.0013779034130452991, + "loss": 3.9675703048706055, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.013464349322021008, + "learning_rate": 0.0013765762281208868, + "loss": 3.9901437759399414, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.009626250714063644, + "learning_rate": 0.0013752491547220063, + "loss": 3.9373281002044678, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.00897257961332798, + "learning_rate": 0.0013739221939178582, + "loss": 4.00804328918457, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.010689414106309414, + "learning_rate": 0.00137259534677755, + "loss": 4.020804405212402, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.010462156496942043, + "learning_rate": 0.0013712686143700988, + "loss": 4.028973579406738, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.009135857224464417, + "learning_rate": 0.001369941997764429, + "loss": 3.9798216819763184, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.009298899210989475, + "learning_rate": 0.0013686154980293713, + "loss": 4.018845081329346, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.00979928020387888, + "learning_rate": 0.0013672891162336636, + "loss": 3.9946775436401367, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.008175542578101158, + "learning_rate": 0.0013659628534459457, + "loss": 3.981858491897583, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.008302642032504082, + "learning_rate": 0.0013646367107347662, + "loss": 3.991128444671631, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.008516009896993637, + "learning_rate": 0.0013633106891685718, + "loss": 4.004990100860596, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.007936482317745686, + "learning_rate": 0.0013619847898157164, + "loss": 3.995734214782715, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.008317600004374981, + "learning_rate": 0.001360659013744451, + "loss": 4.040653228759766, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.008481282740831375, + "learning_rate": 0.001359333362022931, + "loss": 3.969860553741455, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.008944948203861713, + "learning_rate": 0.0013580078357192096, + "loss": 3.9908909797668457, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.008897850289940834, + "learning_rate": 0.001356682435901239, + "loss": 4.04786491394043, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.009527456946671009, + "learning_rate": 0.001355357163636871, + "loss": 4.017423629760742, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.011714582331478596, + "learning_rate": 0.001354032019993852, + "loss": 3.9990692138671875, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.011083807796239853, + "learning_rate": 0.0013527070060398274, + "loss": 3.995265483856201, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.009878288954496384, + "learning_rate": 0.0013513821228423363, + "loss": 4.0364532470703125, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.009727329947054386, + "learning_rate": 0.0013500573714688136, + "loss": 4.035953521728516, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.009482033550739288, + "learning_rate": 0.0013487327529865879, + "loss": 4.015855312347412, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.009037062525749207, + "learning_rate": 0.0013474082684628798, + "loss": 4.024164199829102, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.009094092063605785, + "learning_rate": 0.0013460839189648028, + "loss": 4.037765979766846, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.008230654522776604, + "learning_rate": 0.0013447597055593612, + "loss": 4.006317615509033, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.008491289801895618, + "learning_rate": 0.0013434356293134501, + "loss": 4.042603969573975, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.008907693438231945, + "learning_rate": 0.0013421116912938533, + "loss": 4.008098602294922, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.008897818624973297, + "learning_rate": 0.0013407878925672438, + "loss": 3.9892945289611816, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.00957555789500475, + "learning_rate": 0.0013394642342001832, + "loss": 4.046009063720703, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.009948568418622017, + "learning_rate": 0.0013381407172591185, + "loss": 3.9760642051696777, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.01014955248683691, + "learning_rate": 0.0013368173428103835, + "loss": 4.023215293884277, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.008578895591199398, + "learning_rate": 0.0013354941119201966, + "loss": 4.002819061279297, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.008729875087738037, + "learning_rate": 0.001334171025654662, + "loss": 4.0340657234191895, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.009556280449032784, + "learning_rate": 0.0013328480850797655, + "loss": 3.983076333999634, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.009733820334076881, + "learning_rate": 0.0013315252912613766, + "loss": 4.066578388214111, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.009025252424180508, + "learning_rate": 0.0013302026452652474, + "loss": 3.9980697631835938, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.008485320024192333, + "learning_rate": 0.001328880148157008, + "loss": 4.018826961517334, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.009539696387946606, + "learning_rate": 0.0013275578010021726, + "loss": 4.011737823486328, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.010881016030907631, + "learning_rate": 0.0013262356048661305, + "loss": 4.007291793823242, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.010362236760556698, + "learning_rate": 0.0013249135608141523, + "loss": 3.9929394721984863, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.010296608321368694, + "learning_rate": 0.0013235916699113849, + "loss": 4.0015130043029785, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.011211174540221691, + "learning_rate": 0.0013222699332228517, + "loss": 4.005324363708496, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.009743994101881981, + "learning_rate": 0.0013209483518134525, + "loss": 4.0278143882751465, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.008563914336264133, + "learning_rate": 0.0013196269267479607, + "loss": 3.9997801780700684, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.009078902192413807, + "learning_rate": 0.0013183056590910255, + "loss": 4.026200294494629, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.009618895128369331, + "learning_rate": 0.0013169845499071675, + "loss": 4.027050018310547, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.008920644409954548, + "learning_rate": 0.0013156636002607822, + "loss": 4.000490665435791, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.008849202655255795, + "learning_rate": 0.0013143428112161323, + "loss": 4.044985294342041, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.009303621016442776, + "learning_rate": 0.001313022183837356, + "loss": 4.046281814575195, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.01076227892190218, + "learning_rate": 0.0013117017191884579, + "loss": 4.030576705932617, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.010400841943919659, + "learning_rate": 0.0013103814183333127, + "loss": 3.980081558227539, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.009459166787564754, + "learning_rate": 0.001309061282335663, + "loss": 4.014824867248535, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.009054969064891338, + "learning_rate": 0.001307741312259118, + "loss": 4.0485944747924805, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.009003991261124611, + "learning_rate": 0.0013064215091671545, + "loss": 3.9990320205688477, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.008959448896348476, + "learning_rate": 0.001305101874123113, + "loss": 4.038620948791504, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.009102243930101395, + "learning_rate": 0.0013037824081902015, + "loss": 4.030906677246094, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.008599337190389633, + "learning_rate": 0.0013024631124314877, + "loss": 4.02648401260376, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.008771977387368679, + "learning_rate": 0.0013011439879099056, + "loss": 4.027151107788086, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.009082595817744732, + "learning_rate": 0.00129982503568825, + "loss": 4.0194478034973145, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.009025354869663715, + "learning_rate": 0.0012985062568291763, + "loss": 4.044382095336914, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.00875102635473013, + "learning_rate": 0.0012971876523952014, + "loss": 4.0257568359375, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.009312704205513, + "learning_rate": 0.0012958692234487006, + "loss": 3.9964494705200195, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.431707859039307, + "eval_runtime": 50.7311, + "eval_samples_per_second": 48.136, + "eval_steps_per_second": 1.518, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.00992550328373909, + "learning_rate": 0.0012945509710519088, + "loss": 3.9253792762756348, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.009920453652739525, + "learning_rate": 0.0012932328962669173, + "loss": 3.9438819885253906, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.009437218308448792, + "learning_rate": 0.0012919150001556756, + "loss": 3.960963249206543, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.009616893716156483, + "learning_rate": 0.0012905972837799896, + "loss": 3.969301223754883, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.009107022546231747, + "learning_rate": 0.0012892797482015185, + "loss": 3.9626078605651855, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.011317167431116104, + "learning_rate": 0.0012879623944817772, + "loss": 3.946483850479126, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.010536770336329937, + "learning_rate": 0.001286645223682134, + "loss": 3.959566116333008, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.009349033236503601, + "learning_rate": 0.0012853282368638097, + "loss": 3.964435338973999, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.009950080886483192, + "learning_rate": 0.0012840114350878762, + "loss": 3.9729413986206055, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.010833023115992546, + "learning_rate": 0.001282694819415257, + "loss": 4.006421089172363, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.011538525111973286, + "learning_rate": 0.0012813783909067266, + "loss": 3.9558355808258057, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.010671315714716911, + "learning_rate": 0.0012800621506229065, + "loss": 4.0052337646484375, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.010835555382072926, + "learning_rate": 0.0012787460996242686, + "loss": 3.958641767501831, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.01082367543131113, + "learning_rate": 0.0012774302389711304, + "loss": 3.9818902015686035, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.009194346144795418, + "learning_rate": 0.0012761145697236583, + "loss": 3.9688854217529297, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.007859935984015465, + "learning_rate": 0.001274799092941862, + "loss": 3.9730639457702637, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.008534743450582027, + "learning_rate": 0.0012734838096855978, + "loss": 3.975003719329834, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.009369445964694023, + "learning_rate": 0.0012721687210145666, + "loss": 3.972907304763794, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.0090147340670228, + "learning_rate": 0.0012708538279883098, + "loss": 3.964256525039673, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.007745139300823212, + "learning_rate": 0.0012695391316662151, + "loss": 3.9504833221435547, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.008920101448893547, + "learning_rate": 0.0012682246331075075, + "loss": 3.9735097885131836, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.008987226523458958, + "learning_rate": 0.0012669103333712568, + "loss": 3.9482927322387695, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.00841289572417736, + "learning_rate": 0.0012655962335163685, + "loss": 3.965221881866455, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.008016116917133331, + "learning_rate": 0.0012642823346015902, + "loss": 3.9390206336975098, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.008470152504742146, + "learning_rate": 0.0012629686376855068, + "loss": 3.9257445335388184, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.008637234568595886, + "learning_rate": 0.0012616551438265394, + "loss": 3.9999356269836426, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.008036923594772816, + "learning_rate": 0.0012603418540829473, + "loss": 3.9755172729492188, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.008908843621611595, + "learning_rate": 0.0012590287695128232, + "loss": 3.9248545169830322, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.008743193000555038, + "learning_rate": 0.001257715891174096, + "loss": 3.9619390964508057, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.00834199320524931, + "learning_rate": 0.001256403220124528, + "loss": 3.961569309234619, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.009148893877863884, + "learning_rate": 0.001255090757421715, + "loss": 3.972104549407959, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.00925084762275219, + "learning_rate": 0.0012537785041230843, + "loss": 4.000775337219238, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.008659154176712036, + "learning_rate": 0.0012524664612858947, + "loss": 3.954564332962036, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.007787859998643398, + "learning_rate": 0.0012511546299672355, + "loss": 3.9498860836029053, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.00817918125540018, + "learning_rate": 0.0012498430112240246, + "loss": 4.003897666931152, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.009282548911869526, + "learning_rate": 0.001248531606113011, + "loss": 3.9621925354003906, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.009375610388815403, + "learning_rate": 0.001247220415690769, + "loss": 3.9977524280548096, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.009650378488004208, + "learning_rate": 0.001245909441013701, + "loss": 3.9882829189300537, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.009350919164717197, + "learning_rate": 0.0012445986831380366, + "loss": 4.019596099853516, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.008726303465664387, + "learning_rate": 0.0012432881431198289, + "loss": 3.9933621883392334, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.00829484686255455, + "learning_rate": 0.0012419778220149568, + "loss": 4.001237869262695, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.008744452148675919, + "learning_rate": 0.0012406677208791218, + "loss": 3.9323887825012207, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.008170492947101593, + "learning_rate": 0.0012393578407678496, + "loss": 3.925581455230713, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.007843422703444958, + "learning_rate": 0.001238048182736486, + "loss": 4.001023769378662, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.008914309553802013, + "learning_rate": 0.0012367387478401987, + "loss": 3.9804537296295166, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.009736751206219196, + "learning_rate": 0.0012354295371339774, + "loss": 3.9831767082214355, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.010061381384730339, + "learning_rate": 0.0012341205516726273, + "loss": 3.9462435245513916, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.010350692085921764, + "learning_rate": 0.0012328117925107765, + "loss": 4.012863636016846, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01039902213960886, + "learning_rate": 0.0012315032607028667, + "loss": 3.9811301231384277, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.011423371732234955, + "learning_rate": 0.0012301949573031595, + "loss": 3.9773542881011963, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.012449614703655243, + "learning_rate": 0.001228886883365731, + "loss": 3.969600200653076, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.010361429303884506, + "learning_rate": 0.0012275790399444726, + "loss": 4.035360336303711, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.009603082202374935, + "learning_rate": 0.0012262714280930904, + "loss": 3.9840846061706543, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.009183845482766628, + "learning_rate": 0.0012249640488651036, + "loss": 4.031542778015137, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.009694446809589863, + "learning_rate": 0.0012236569033138441, + "loss": 3.9935057163238525, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.0097816726192832, + "learning_rate": 0.0012223499924924545, + "loss": 3.9851176738739014, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.009413184598088264, + "learning_rate": 0.001221043317453891, + "loss": 3.9642205238342285, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.009960711002349854, + "learning_rate": 0.0012197368792509166, + "loss": 3.9373817443847656, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.010124670341610909, + "learning_rate": 0.0012184306789361059, + "loss": 3.958115577697754, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.011087656021118164, + "learning_rate": 0.00121712471756184, + "loss": 3.9607057571411133, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.009812800213694572, + "learning_rate": 0.001215818996180309, + "loss": 4.00722599029541, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.010305573232471943, + "learning_rate": 0.0012145135158435094, + "loss": 3.98728084564209, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.01019787136465311, + "learning_rate": 0.0012132082776032422, + "loss": 3.981172800064087, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.009260181337594986, + "learning_rate": 0.001211903282511115, + "loss": 4.034339904785156, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.00972998607903719, + "learning_rate": 0.001210598531618538, + "loss": 4.022885322570801, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.009853586554527283, + "learning_rate": 0.0012092940259767267, + "loss": 4.000235080718994, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.00961449183523655, + "learning_rate": 0.0012079897666366962, + "loss": 3.9951353073120117, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.010257176123559475, + "learning_rate": 0.0012066857546492657, + "loss": 3.9801371097564697, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.009443561546504498, + "learning_rate": 0.001205381991065054, + "loss": 3.9596502780914307, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.010389018803834915, + "learning_rate": 0.0012040784769344795, + "loss": 3.980527639389038, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.010322567075490952, + "learning_rate": 0.0012027752133077607, + "loss": 3.996258497238159, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.009987728670239449, + "learning_rate": 0.0012014722012349125, + "loss": 3.9493203163146973, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.421086311340332, + "eval_runtime": 52.5623, + "eval_samples_per_second": 46.459, + "eval_steps_per_second": 1.465, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.009405115619301796, + "learning_rate": 0.0012001694417657493, + "loss": 3.927506923675537, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.009974643588066101, + "learning_rate": 0.00119886693594988, + "loss": 3.9373550415039062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.011331096291542053, + "learning_rate": 0.0011975646848367098, + "loss": 3.935594081878662, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.010538448579609394, + "learning_rate": 0.0011962626894754405, + "loss": 3.911579132080078, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.00989572424441576, + "learning_rate": 0.0011949609509150645, + "loss": 3.92525577545166, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.010533696971833706, + "learning_rate": 0.0011936594702043704, + "loss": 3.898075580596924, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.01051455456763506, + "learning_rate": 0.0011923582483919365, + "loss": 3.93463134765625, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.009662075899541378, + "learning_rate": 0.0011910572865261348, + "loss": 3.898221731185913, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.008874096907675266, + "learning_rate": 0.0011897565856551257, + "loss": 3.926787853240967, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.008126807399094105, + "learning_rate": 0.0011884561468268606, + "loss": 3.9158413410186768, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.007915553636848927, + "learning_rate": 0.001187155971089081, + "loss": 3.8891496658325195, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.008312922902405262, + "learning_rate": 0.0011858560594893129, + "loss": 3.932219982147217, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.008892328478395939, + "learning_rate": 0.0011845564130748734, + "loss": 3.8866336345672607, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.009187526069581509, + "learning_rate": 0.0011832570328928628, + "loss": 3.9464707374572754, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.008615494705736637, + "learning_rate": 0.0011819579199901693, + "loss": 3.9396629333496094, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.008649308234453201, + "learning_rate": 0.0011806590754134636, + "loss": 3.9325060844421387, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.008581082336604595, + "learning_rate": 0.0011793605002092021, + "loss": 3.9200944900512695, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.008502699434757233, + "learning_rate": 0.0011780621954236238, + "loss": 3.9656100273132324, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.0083576999604702, + "learning_rate": 0.0011767641621027477, + "loss": 3.9130215644836426, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.008196269161999226, + "learning_rate": 0.0011754664012923785, + "loss": 3.942369222640991, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.008857307955622673, + "learning_rate": 0.0011741689140380956, + "loss": 3.909585952758789, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.008935349993407726, + "learning_rate": 0.0011728717013852635, + "loss": 3.9248366355895996, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.008821200579404831, + "learning_rate": 0.0011715747643790211, + "loss": 3.9589288234710693, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.00861942209303379, + "learning_rate": 0.0011702781040642885, + "loss": 3.9381728172302246, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.009127574972808361, + "learning_rate": 0.0011689817214857613, + "loss": 3.897829532623291, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.008008269593119621, + "learning_rate": 0.0011676856176879107, + "loss": 3.930056095123291, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.010070208460092545, + "learning_rate": 0.0011663897937149853, + "loss": 3.892270803451538, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.00903113093227148, + "learning_rate": 0.001165094250611006, + "loss": 3.9766440391540527, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.008872415870428085, + "learning_rate": 0.0011637989894197692, + "loss": 3.9559712409973145, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.009524816647171974, + "learning_rate": 0.0011625040111848427, + "loss": 3.9607882499694824, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.009836779907345772, + "learning_rate": 0.0011612093169495678, + "loss": 3.9629592895507812, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.009586540050804615, + "learning_rate": 0.0011599149077570561, + "loss": 3.95208740234375, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.010495901107788086, + "learning_rate": 0.0011586207846501896, + "loss": 3.9439730644226074, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.010346266441047192, + "learning_rate": 0.0011573269486716202, + "loss": 3.959299087524414, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.010767634026706219, + "learning_rate": 0.0011560334008637674, + "loss": 3.900784969329834, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.012122200801968575, + "learning_rate": 0.0011547401422688202, + "loss": 3.9588193893432617, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.012914645485579967, + "learning_rate": 0.0011534471739287331, + "loss": 3.9278407096862793, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.011785333044826984, + "learning_rate": 0.0011521544968852273, + "loss": 3.992793560028076, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.008455781266093254, + "learning_rate": 0.0011508621121797899, + "loss": 3.96525239944458, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.009655282832682133, + "learning_rate": 0.0011495700208536714, + "loss": 3.9346890449523926, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.010537004098296165, + "learning_rate": 0.0011482782239478871, + "loss": 3.9534668922424316, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.009438651613891125, + "learning_rate": 0.0011469867225032135, + "loss": 3.983487606048584, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.010586646385490894, + "learning_rate": 0.0011456955175601908, + "loss": 3.939680576324463, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.009082718752324581, + "learning_rate": 0.001144404610159119, + "loss": 3.9558515548706055, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.009833602234721184, + "learning_rate": 0.0011431140013400591, + "loss": 3.9539828300476074, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.009142170660197735, + "learning_rate": 0.0011418236921428313, + "loss": 3.9442849159240723, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.008603853173553944, + "learning_rate": 0.0011405336836070137, + "loss": 3.9526913166046143, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.008838302455842495, + "learning_rate": 0.0011392439767719446, + "loss": 3.973104953765869, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.008867006748914719, + "learning_rate": 0.0011379545726767156, + "loss": 3.9832305908203125, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.00971697736531496, + "learning_rate": 0.001136665472360178, + "loss": 3.9563937187194824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.00957175437361002, + "learning_rate": 0.001135376676860935, + "loss": 3.975194215774536, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.008548660203814507, + "learning_rate": 0.001134088187217347, + "loss": 3.9839985370635986, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.008556334301829338, + "learning_rate": 0.0011328000044675264, + "loss": 3.9225516319274902, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.008440867997705936, + "learning_rate": 0.0011315121296493385, + "loss": 3.945028066635132, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.009643333032727242, + "learning_rate": 0.0011302245638004017, + "loss": 3.943120002746582, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.008832835592329502, + "learning_rate": 0.001128937307958083, + "loss": 3.9895920753479004, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.008720474317669868, + "learning_rate": 0.0011276503631595036, + "loss": 3.941950798034668, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.008099926635622978, + "learning_rate": 0.001126363730441529, + "loss": 3.9626870155334473, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.008525107055902481, + "learning_rate": 0.0011250774108407778, + "loss": 3.950535774230957, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.008998272009193897, + "learning_rate": 0.0011237914053936135, + "loss": 3.9460678100585938, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.009387437254190445, + "learning_rate": 0.001122505715136148, + "loss": 3.926604986190796, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.007804017513990402, + "learning_rate": 0.0011212203411042386, + "loss": 3.9894230365753174, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.008564616553485394, + "learning_rate": 0.0011199352843334876, + "loss": 3.9339959621429443, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.008726440370082855, + "learning_rate": 0.0011186505458592427, + "loss": 3.9867987632751465, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.008074100129306316, + "learning_rate": 0.0011173661267165932, + "loss": 3.9492292404174805, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.007688041776418686, + "learning_rate": 0.0011160820279403744, + "loss": 3.9425454139709473, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.007808993570506573, + "learning_rate": 0.0011147982505651595, + "loss": 3.935114860534668, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.009786820039153099, + "learning_rate": 0.0011135147956252659, + "loss": 3.959758758544922, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.009487155824899673, + "learning_rate": 0.00111223166415475, + "loss": 3.989048480987549, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.008407804183661938, + "learning_rate": 0.0011109488571874073, + "loss": 3.932347297668457, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.007158380467444658, + "learning_rate": 0.0011096663757567723, + "loss": 3.9545154571533203, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.008840581402182579, + "learning_rate": 0.001108384220896117, + "loss": 3.965984344482422, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.409194469451904, + "eval_runtime": 51.5561, + "eval_samples_per_second": 47.366, + "eval_steps_per_second": 1.494, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.010702896863222122, + "learning_rate": 0.0011071023936384508, + "loss": 3.8964462280273438, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01290619745850563, + "learning_rate": 0.0011058208950165183, + "loss": 3.8775625228881836, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01443667896091938, + "learning_rate": 0.0011045397260628, + "loss": 3.8755288124084473, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.014411968179047108, + "learning_rate": 0.0011032588878095108, + "loss": 3.889918327331543, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.013656266033649445, + "learning_rate": 0.0011019783812885986, + "loss": 3.926997661590576, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.010661937296390533, + "learning_rate": 0.0011006982075317446, + "loss": 3.9226956367492676, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.008968768641352654, + "learning_rate": 0.0010994183675703615, + "loss": 3.8932647705078125, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.009992181323468685, + "learning_rate": 0.0010981388624355937, + "loss": 3.8920068740844727, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.011034460738301277, + "learning_rate": 0.0010968596931583148, + "loss": 3.909073829650879, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.011815821751952171, + "learning_rate": 0.0010955808607691284, + "loss": 3.9190006256103516, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.012317490763962269, + "learning_rate": 0.0010943023662983674, + "loss": 3.8790552616119385, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.010983958840370178, + "learning_rate": 0.0010930242107760903, + "loss": 3.8730852603912354, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.008948802947998047, + "learning_rate": 0.0010917463952320856, + "loss": 3.923659563064575, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.009568693116307259, + "learning_rate": 0.0010904689206958649, + "loss": 3.9100794792175293, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.011862963438034058, + "learning_rate": 0.0010891917881966672, + "loss": 3.908093214035034, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.010180995799601078, + "learning_rate": 0.0010879149987634545, + "loss": 3.900660514831543, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.009372464381158352, + "learning_rate": 0.0010866385534249131, + "loss": 3.8950703144073486, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.010695629753172398, + "learning_rate": 0.0010853624532094525, + "loss": 3.913530111312866, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.011817459017038345, + "learning_rate": 0.0010840866991452024, + "loss": 3.90268611907959, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.010686337016522884, + "learning_rate": 0.0010828112922600167, + "loss": 3.9141364097595215, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.009550466202199459, + "learning_rate": 0.0010815362335814655, + "loss": 3.968468189239502, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.009367208927869797, + "learning_rate": 0.0010802615241368423, + "loss": 3.8933968544006348, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.009899397380650043, + "learning_rate": 0.0010789871649531562, + "loss": 3.936708450317383, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.009181005880236626, + "learning_rate": 0.0010777131570571363, + "loss": 3.9370503425598145, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.00799303874373436, + "learning_rate": 0.0010764395014752274, + "loss": 3.911050319671631, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.008469315245747566, + "learning_rate": 0.0010751661992335908, + "loss": 3.9103240966796875, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.009814459830522537, + "learning_rate": 0.0010738932513581034, + "loss": 3.881843090057373, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.010244108736515045, + "learning_rate": 0.0010726206588743557, + "loss": 3.883141040802002, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.008898732252418995, + "learning_rate": 0.001071348422807653, + "loss": 3.952347755432129, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.00823870487511158, + "learning_rate": 0.0010700765441830121, + "loss": 3.932507038116455, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.009857657365500927, + "learning_rate": 0.001068805024025164, + "loss": 3.932932138442993, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.00929308496415615, + "learning_rate": 0.0010675338633585488, + "loss": 3.8971338272094727, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.008667961694300175, + "learning_rate": 0.0010662630632073172, + "loss": 3.916931629180908, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.009758287109434605, + "learning_rate": 0.001064992624595331, + "loss": 3.964726209640503, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.009429254569113255, + "learning_rate": 0.0010637225485461586, + "loss": 3.917572498321533, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.008303353562951088, + "learning_rate": 0.0010624528360830782, + "loss": 3.875241756439209, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.008729013614356518, + "learning_rate": 0.001061183488229073, + "loss": 3.9691600799560547, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.010239302180707455, + "learning_rate": 0.0010599145060068353, + "loss": 3.9053149223327637, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.00932721421122551, + "learning_rate": 0.0010586458904387596, + "loss": 3.8593955039978027, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.008196988143026829, + "learning_rate": 0.0010573776425469473, + "loss": 3.917983055114746, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.008049141615629196, + "learning_rate": 0.001056109763353203, + "loss": 3.9337785243988037, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.009256313554942608, + "learning_rate": 0.001054842253879034, + "loss": 3.937887668609619, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.009292179718613625, + "learning_rate": 0.0010535751151456495, + "loss": 3.9203450679779053, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.009142161346971989, + "learning_rate": 0.0010523083481739602, + "loss": 3.9027457237243652, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.009498288854956627, + "learning_rate": 0.0010510419539845777, + "loss": 3.917316436767578, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.00849096104502678, + "learning_rate": 0.001049775933597813, + "loss": 3.9229235649108887, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.008343606255948544, + "learning_rate": 0.0010485102880336747, + "loss": 3.9138073921203613, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.010342134162783623, + "learning_rate": 0.0010472450183118727, + "loss": 3.8932929039001465, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.010298306122422218, + "learning_rate": 0.00104598012545181, + "loss": 3.9112677574157715, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.00827560480684042, + "learning_rate": 0.001044715610472589, + "loss": 3.894556760787964, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.008993102237582207, + "learning_rate": 0.0010434514743930064, + "loss": 3.905391216278076, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.008632538840174675, + "learning_rate": 0.0010421877182315536, + "loss": 3.896714210510254, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.008542424067854881, + "learning_rate": 0.0010409243430064162, + "loss": 3.9145569801330566, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.008970450609922409, + "learning_rate": 0.0010396613497354725, + "loss": 3.9170360565185547, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.008744142949581146, + "learning_rate": 0.0010383987394362942, + "loss": 3.927412509918213, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.008229847066104412, + "learning_rate": 0.0010371365131261424, + "loss": 3.928352117538452, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.00818693358451128, + "learning_rate": 0.0010358746718219717, + "loss": 3.9219343662261963, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.008848588913679123, + "learning_rate": 0.001034613216540423, + "loss": 3.9346699714660645, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.00834780652076006, + "learning_rate": 0.0010333521482978295, + "loss": 3.9144492149353027, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.00940527394413948, + "learning_rate": 0.0010320914681102098, + "loss": 3.951573371887207, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.008884521201252937, + "learning_rate": 0.0010308311769932725, + "loss": 3.899625778198242, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.007863261736929417, + "learning_rate": 0.0010295712759624107, + "loss": 3.9321727752685547, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.008651153184473515, + "learning_rate": 0.0010283117660327035, + "loss": 3.9030094146728516, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.00822074618190527, + "learning_rate": 0.0010270526482189163, + "loss": 3.939466953277588, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.008701153099536896, + "learning_rate": 0.0010257939235354964, + "loss": 3.9099626541137695, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.008418620564043522, + "learning_rate": 0.0010245355929965774, + "loss": 3.914116382598877, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.008339700289070606, + "learning_rate": 0.001023277657615971, + "loss": 3.9111878871917725, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.008304647170007229, + "learning_rate": 0.0010220201184071747, + "loss": 3.950974464416504, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.008436229079961777, + "learning_rate": 0.001020762976383365, + "loss": 3.9631457328796387, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.0074487063102424145, + "learning_rate": 0.001019506232557398, + "loss": 3.902984619140625, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.008736515417695045, + "learning_rate": 0.0010182498879418096, + "loss": 3.9268949031829834, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.009666468016803265, + "learning_rate": 0.0010169939435488145, + "loss": 3.9274301528930664, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.395970821380615, + "eval_runtime": 53.7848, + "eval_samples_per_second": 45.403, + "eval_steps_per_second": 1.432, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.009094545617699623, + "learning_rate": 0.001015738400390304, + "loss": 3.866425037384033, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.009273733012378216, + "learning_rate": 0.0010144832594778462, + "loss": 3.8569841384887695, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.009205271489918232, + "learning_rate": 0.0010132285218226855, + "loss": 3.8981378078460693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.009963045828044415, + "learning_rate": 0.0010119741884357424, + "loss": 3.881542444229126, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.008536554872989655, + "learning_rate": 0.0010107202603276096, + "loss": 3.851253032684326, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.01074561569839716, + "learning_rate": 0.001009466738508555, + "loss": 3.8731184005737305, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.010809902101755142, + "learning_rate": 0.0010082136239885177, + "loss": 3.863889217376709, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.008289352990686893, + "learning_rate": 0.0010069609177771103, + "loss": 3.839066743850708, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.00792208593338728, + "learning_rate": 0.001005708620883615, + "loss": 3.8634843826293945, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.008381598629057407, + "learning_rate": 0.0010044567343169845, + "loss": 3.868283987045288, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.009784617461264133, + "learning_rate": 0.0010032052590858425, + "loss": 3.853074550628662, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.009462198242545128, + "learning_rate": 0.0010019541961984782, + "loss": 3.889979600906372, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.010207819752395153, + "learning_rate": 0.001000703546662852, + "loss": 3.8636741638183594, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.009123551659286022, + "learning_rate": 0.0009994533114865887, + "loss": 3.8578555583953857, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.009006140753626823, + "learning_rate": 0.0009982034916769806, + "loss": 3.8828015327453613, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.0091114342212677, + "learning_rate": 0.0009969540882409846, + "loss": 3.880486011505127, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.008851367980241776, + "learning_rate": 0.0009957051021852227, + "loss": 3.8618125915527344, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.008196741342544556, + "learning_rate": 0.0009944565345159806, + "loss": 3.8314175605773926, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.00863513071089983, + "learning_rate": 0.000993208386239206, + "loss": 3.856212615966797, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.008624681271612644, + "learning_rate": 0.0009919606583605107, + "loss": 3.868927001953125, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.009208842180669308, + "learning_rate": 0.0009907133518851647, + "loss": 3.9072189331054688, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.010200913064181805, + "learning_rate": 0.0009894664678181027, + "loss": 3.875272750854492, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.009892710484564304, + "learning_rate": 0.000988220007163914, + "loss": 3.888686418533325, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.009873585775494576, + "learning_rate": 0.000986973970926851, + "loss": 3.8607118129730225, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.009322221390902996, + "learning_rate": 0.0009857283601108228, + "loss": 3.8852052688598633, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.008141528815031052, + "learning_rate": 0.0009844831757193947, + "loss": 3.888213872909546, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.008191613480448723, + "learning_rate": 0.00098323841875579, + "loss": 3.9073967933654785, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.008419081568717957, + "learning_rate": 0.0009819940902228864, + "loss": 3.879223346710205, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.009037869982421398, + "learning_rate": 0.0009807501911232171, + "loss": 3.883303165435791, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.00854078121483326, + "learning_rate": 0.0009795067224589687, + "loss": 3.926767587661743, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.008590771816670895, + "learning_rate": 0.0009782636852319825, + "loss": 3.913999557495117, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.00859815627336502, + "learning_rate": 0.0009770210804437498, + "loss": 3.8828063011169434, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.008970260620117188, + "learning_rate": 0.0009757789090954157, + "loss": 3.9174866676330566, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.00877175759524107, + "learning_rate": 0.0009745371721877754, + "loss": 3.8879401683807373, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.008541764691472054, + "learning_rate": 0.0009732958707212731, + "loss": 3.8717427253723145, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.008466241881251335, + "learning_rate": 0.0009720550056960039, + "loss": 3.8944485187530518, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.00808491837233305, + "learning_rate": 0.0009708145781117096, + "loss": 3.8617026805877686, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.00842207670211792, + "learning_rate": 0.0009695745889677805, + "loss": 3.887824535369873, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.00830599945038557, + "learning_rate": 0.0009683350392632533, + "loss": 3.8612070083618164, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.008417914621531963, + "learning_rate": 0.0009670959299968111, + "loss": 3.880140781402588, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.008542059920728207, + "learning_rate": 0.0009658572621667822, + "loss": 3.8951029777526855, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.008822236210107803, + "learning_rate": 0.000964619036771138, + "loss": 3.8962583541870117, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.009151184931397438, + "learning_rate": 0.0009633812548074952, + "loss": 3.882500648498535, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.007803036365658045, + "learning_rate": 0.0009621439172731115, + "loss": 3.8740782737731934, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.009350850246846676, + "learning_rate": 0.0009609070251648884, + "loss": 3.8627419471740723, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.008319144137203693, + "learning_rate": 0.0009596705794793667, + "loss": 3.9084486961364746, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.009198528714478016, + "learning_rate": 0.0009584345812127282, + "loss": 3.92141056060791, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.010480938479304314, + "learning_rate": 0.0009571990313607958, + "loss": 3.8916103839874268, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.01045288611203432, + "learning_rate": 0.0009559639309190278, + "loss": 3.8863210678100586, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.009121391922235489, + "learning_rate": 0.0009547292808825239, + "loss": 3.9167044162750244, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.01093223039060831, + "learning_rate": 0.0009534950822460184, + "loss": 3.9046711921691895, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.01214868575334549, + "learning_rate": 0.0009522613360038835, + "loss": 3.9126768112182617, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.01055280864238739, + "learning_rate": 0.0009510280431501255, + "loss": 3.9214353561401367, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.009021313861012459, + "learning_rate": 0.0009497952046783867, + "loss": 3.8795580863952637, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.00863752793520689, + "learning_rate": 0.0009485628215819428, + "loss": 3.893232822418213, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.009327626787126064, + "learning_rate": 0.0009473308948537021, + "loss": 3.917377471923828, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.01003898587077856, + "learning_rate": 0.0009460994254862066, + "loss": 3.8818931579589844, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.00892532430589199, + "learning_rate": 0.0009448684144716278, + "loss": 3.8761978149414062, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.008396463468670845, + "learning_rate": 0.0009436378628017702, + "loss": 3.8665928840637207, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.00784370768815279, + "learning_rate": 0.0009424077714680662, + "loss": 3.8850138187408447, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.00809321366250515, + "learning_rate": 0.0009411781414615784, + "loss": 3.917078733444214, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.008685288950800896, + "learning_rate": 0.000939948973772998, + "loss": 3.8865013122558594, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.007291820365935564, + "learning_rate": 0.0009387202693926425, + "loss": 3.8967666625976562, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.008459319360554218, + "learning_rate": 0.0009374920293104573, + "loss": 3.9043235778808594, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.007924137637019157, + "learning_rate": 0.0009362642545160123, + "loss": 3.9216856956481934, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.007874754257500172, + "learning_rate": 0.0009350369459985052, + "loss": 3.879725933074951, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.008643723092973232, + "learning_rate": 0.0009338101047467544, + "loss": 3.9303886890411377, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.008451704867184162, + "learning_rate": 0.000932583731749205, + "loss": 3.9352827072143555, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.008137754164636135, + "learning_rate": 0.0009313578279939236, + "loss": 3.930800437927246, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.009277567267417908, + "learning_rate": 0.0009301323944685979, + "loss": 3.9194300174713135, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.009428970515727997, + "learning_rate": 0.0009289074321605383, + "loss": 3.874298572540283, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.0089727733284235, + "learning_rate": 0.0009276829420566743, + "loss": 3.893770694732666, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.388528823852539, + "eval_runtime": 51.5859, + "eval_samples_per_second": 47.338, + "eval_steps_per_second": 1.493, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.009470812045037746, + "learning_rate": 0.000926458925143556, + "loss": 3.8514318466186523, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.011485518887639046, + "learning_rate": 0.0009252353824073511, + "loss": 3.857734203338623, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.010095939971506596, + "learning_rate": 0.0009240123148338457, + "loss": 3.819117546081543, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.010103728622198105, + "learning_rate": 0.0009227897234084444, + "loss": 3.842660427093506, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.01148147787898779, + "learning_rate": 0.0009215676091161661, + "loss": 3.821089029312134, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.012331553734838963, + "learning_rate": 0.0009203459729416467, + "loss": 3.8194339275360107, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.011274794116616249, + "learning_rate": 0.0009191248158691358, + "loss": 3.8839240074157715, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.009603365324437618, + "learning_rate": 0.0009179041388824986, + "loss": 3.806617259979248, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.010824795812368393, + "learning_rate": 0.0009166839429652112, + "loss": 3.823228597640991, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.01014280878007412, + "learning_rate": 0.0009154642291003637, + "loss": 3.860605239868164, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.009052124805748463, + "learning_rate": 0.0009142449982706591, + "loss": 3.84226131439209, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.009465130977332592, + "learning_rate": 0.0009130262514584074, + "loss": 3.8674912452697754, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.009991783648729324, + "learning_rate": 0.0009118079896455326, + "loss": 3.8775739669799805, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.009288253262639046, + "learning_rate": 0.0009105902138135652, + "loss": 3.7978098392486572, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.009964323602616787, + "learning_rate": 0.0009093729249436459, + "loss": 3.8362321853637695, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.010977157391607761, + "learning_rate": 0.0009081561240165218, + "loss": 3.857304096221924, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.009241319261491299, + "learning_rate": 0.0009069398120125479, + "loss": 3.8659310340881348, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.009211038239300251, + "learning_rate": 0.0009057239899116844, + "loss": 3.8764538764953613, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.009649050422012806, + "learning_rate": 0.000904508658693497, + "loss": 3.8702778816223145, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.00954380165785551, + "learning_rate": 0.0009032938193371573, + "loss": 3.8273327350616455, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.008831270970404148, + "learning_rate": 0.0009020794728214378, + "loss": 3.821664810180664, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.008367286063730717, + "learning_rate": 0.000900865620124717, + "loss": 3.8586034774780273, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.008805174380540848, + "learning_rate": 0.0008996522622249732, + "loss": 3.811522960662842, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.00873226672410965, + "learning_rate": 0.0008984394000997873, + "loss": 3.8538150787353516, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.008930203504860401, + "learning_rate": 0.0008972270347263402, + "loss": 3.8448486328125, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.009300434030592442, + "learning_rate": 0.0008960151670814125, + "loss": 3.842113494873047, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.009598663076758385, + "learning_rate": 0.0008948037981413855, + "loss": 3.886852741241455, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.008400370366871357, + "learning_rate": 0.0008935929288822357, + "loss": 3.852076292037964, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.00891344528645277, + "learning_rate": 0.0008923825602795391, + "loss": 3.8480606079101562, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.00869396049529314, + "learning_rate": 0.0008911726933084678, + "loss": 3.8164706230163574, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.007914266549050808, + "learning_rate": 0.0008899633289437904, + "loss": 3.8768272399902344, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.008398533798754215, + "learning_rate": 0.0008887544681598687, + "loss": 3.8625502586364746, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.00887146033346653, + "learning_rate": 0.0008875461119306601, + "loss": 3.8934593200683594, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.008642833679914474, + "learning_rate": 0.0008863382612297172, + "loss": 3.8391976356506348, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.008099788799881935, + "learning_rate": 0.0008851309170301816, + "loss": 3.827965497970581, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.008166342042386532, + "learning_rate": 0.0008839240803047894, + "loss": 3.8999152183532715, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.009095479734241962, + "learning_rate": 0.0008827177520258671, + "loss": 3.876927614212036, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.009246554225683212, + "learning_rate": 0.0008815119331653321, + "loss": 3.8692450523376465, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.007059223018586636, + "learning_rate": 0.0008803066246946898, + "loss": 3.869400978088379, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.008499744348227978, + "learning_rate": 0.0008791018275850363, + "loss": 3.8303189277648926, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.007667839527130127, + "learning_rate": 0.0008778975428070543, + "loss": 3.827998638153076, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.00765089550986886, + "learning_rate": 0.0008766937713310145, + "loss": 3.8554482460021973, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.00862938817590475, + "learning_rate": 0.0008754905141267749, + "loss": 3.905458927154541, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.008733712136745453, + "learning_rate": 0.0008742877721637758, + "loss": 3.8564133644104004, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.0071738120168447495, + "learning_rate": 0.0008730855464110471, + "loss": 3.878807544708252, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.008418241515755653, + "learning_rate": 0.0008718838378371989, + "loss": 3.8399720191955566, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.008526502177119255, + "learning_rate": 0.0008706826474104263, + "loss": 3.847656488418579, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.00732462527230382, + "learning_rate": 0.0008694819760985069, + "loss": 3.8839492797851562, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.008424196392297745, + "learning_rate": 0.0008682818248688003, + "loss": 3.861239433288574, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.008984955959022045, + "learning_rate": 0.0008670821946882465, + "loss": 3.856529712677002, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.009004034101963043, + "learning_rate": 0.000865883086523365, + "loss": 3.8514528274536133, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.009097260423004627, + "learning_rate": 0.0008646845013402563, + "loss": 3.8890581130981445, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.009244748391211033, + "learning_rate": 0.0008634864401045988, + "loss": 3.872745990753174, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.00788815412670374, + "learning_rate": 0.0008622889037816488, + "loss": 3.8777575492858887, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.008943610824644566, + "learning_rate": 0.0008610918933362396, + "loss": 3.8825736045837402, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.008154108189046383, + "learning_rate": 0.0008598954097327809, + "loss": 3.877514362335205, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.008689315989613533, + "learning_rate": 0.0008586994539352587, + "loss": 3.854720115661621, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.00904505793005228, + "learning_rate": 0.0008575040269072315, + "loss": 3.911463499069214, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.007773210760205984, + "learning_rate": 0.0008563091296118342, + "loss": 3.875370740890503, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.008437697775661945, + "learning_rate": 0.0008551147630117737, + "loss": 3.8457469940185547, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.009780957363545895, + "learning_rate": 0.0008539209280693296, + "loss": 3.890479564666748, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.007929213345050812, + "learning_rate": 0.0008527276257463532, + "loss": 3.876030445098877, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.009121386334300041, + "learning_rate": 0.0008515348570042669, + "loss": 3.9252734184265137, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.00981531199067831, + "learning_rate": 0.000850342622804063, + "loss": 3.8997983932495117, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.009150844067335129, + "learning_rate": 0.0008491509241063025, + "loss": 3.85115385055542, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.00992978923022747, + "learning_rate": 0.0008479597618711154, + "loss": 3.9138102531433105, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.010906689800322056, + "learning_rate": 0.0008467691370582001, + "loss": 3.892329692840576, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.008644519373774529, + "learning_rate": 0.0008455790506268213, + "loss": 3.9071311950683594, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.009030954912304878, + "learning_rate": 0.0008443895035358111, + "loss": 3.867384910583496, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.010806960985064507, + "learning_rate": 0.0008432004967435638, + "loss": 3.877983331680298, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.011148087680339813, + "learning_rate": 0.0008420120312080436, + "loss": 3.8795924186706543, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01046178862452507, + "learning_rate": 0.0008408241078867731, + "loss": 3.887709617614746, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.383835792541504, + "eval_runtime": 53.0778, + "eval_samples_per_second": 46.008, + "eval_steps_per_second": 1.451, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.009721867740154266, + "learning_rate": 0.0008396367277368421, + "loss": 3.8261327743530273, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.01091420091688633, + "learning_rate": 0.000838449891714901, + "loss": 3.7837276458740234, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.010423852130770683, + "learning_rate": 0.0008372636007771623, + "loss": 3.796450138092041, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.011928882449865341, + "learning_rate": 0.0008360778558793996, + "loss": 3.794717311859131, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.010779444128274918, + "learning_rate": 0.000834892657976945, + "loss": 3.8477983474731445, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.009554282762110233, + "learning_rate": 0.0008337080080246915, + "loss": 3.793790817260742, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.009768152609467506, + "learning_rate": 0.0008325239069770906, + "loss": 3.860384225845337, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.010801832191646099, + "learning_rate": 0.0008313403557881513, + "loss": 3.816796064376831, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.01046004518866539, + "learning_rate": 0.0008301573554114378, + "loss": 3.8064301013946533, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.009319447912275791, + "learning_rate": 0.0008289749068000744, + "loss": 3.832408905029297, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01011407095938921, + "learning_rate": 0.000827793010906737, + "loss": 3.8276009559631348, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.010393989272415638, + "learning_rate": 0.0008266116686836582, + "loss": 3.83754825592041, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.00883004255592823, + "learning_rate": 0.0008254308810826244, + "loss": 3.8315513134002686, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.008549787104129791, + "learning_rate": 0.0008242506490549745, + "loss": 3.792919158935547, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.009668390266597271, + "learning_rate": 0.0008230709735516006, + "loss": 3.8110954761505127, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.009008699096739292, + "learning_rate": 0.0008218918555229445, + "loss": 3.8598508834838867, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.008660725317895412, + "learning_rate": 0.0008207132959190026, + "loss": 3.839632511138916, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.00827972125262022, + "learning_rate": 0.0008195352956893172, + "loss": 3.8428683280944824, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.008117233403027058, + "learning_rate": 0.0008183578557829825, + "loss": 3.8434956073760986, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.008314455859363079, + "learning_rate": 0.0008171809771486405, + "loss": 3.8079686164855957, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.00865850131958723, + "learning_rate": 0.0008160046607344811, + "loss": 3.846999406814575, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0096864253282547, + "learning_rate": 0.0008148289074882416, + "loss": 3.8227243423461914, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.00840236246585846, + "learning_rate": 0.0008136537183572045, + "loss": 3.8400344848632812, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.007451791316270828, + "learning_rate": 0.0008124790942881984, + "loss": 3.8318095207214355, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.00858070608228445, + "learning_rate": 0.000811305036227597, + "loss": 3.8232922554016113, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.00765061192214489, + "learning_rate": 0.0008101315451213178, + "loss": 3.81624174118042, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.009304090403020382, + "learning_rate": 0.0008089586219148213, + "loss": 3.8715004920959473, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.009136030450463295, + "learning_rate": 0.0008077862675531103, + "loss": 3.825037956237793, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.007460939697921276, + "learning_rate": 0.0008066144829807303, + "loss": 3.811749219894409, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.007510480936616659, + "learning_rate": 0.0008054432691417659, + "loss": 3.8313679695129395, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.008825462311506271, + "learning_rate": 0.0008042726269798433, + "loss": 3.83119535446167, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.009002318605780602, + "learning_rate": 0.0008031025574381281, + "loss": 3.8542468547821045, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.009266122244298458, + "learning_rate": 0.0008019330614593237, + "loss": 3.8328065872192383, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.008935115300118923, + "learning_rate": 0.0008007641399856727, + "loss": 3.8198719024658203, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.008021850138902664, + "learning_rate": 0.0007995957939589526, + "loss": 3.828382968902588, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.008035171777009964, + "learning_rate": 0.0007984280243204805, + "loss": 3.879868507385254, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.008553579449653625, + "learning_rate": 0.0007972608320111059, + "loss": 3.842162609100342, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.008016746491193771, + "learning_rate": 0.000796094217971215, + "loss": 3.85674786567688, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.007781418971717358, + "learning_rate": 0.0007949281831407276, + "loss": 3.7968859672546387, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.008552255108952522, + "learning_rate": 0.0007937627284590971, + "loss": 3.8305506706237793, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.007390565704554319, + "learning_rate": 0.0007925978548653096, + "loss": 3.8228843212127686, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.007611850742250681, + "learning_rate": 0.0007914335632978812, + "loss": 3.868957996368408, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.008188549429178238, + "learning_rate": 0.0007902698546948628, + "loss": 3.836078405380249, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.008198797702789307, + "learning_rate": 0.0007891067299938318, + "loss": 3.8637044429779053, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.008353965356945992, + "learning_rate": 0.0007879441901318977, + "loss": 3.833843946456909, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.0086967209354043, + "learning_rate": 0.0007867822360456963, + "loss": 3.850536823272705, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.009151408448815346, + "learning_rate": 0.0007856208686713949, + "loss": 3.8453407287597656, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.008010942488908768, + "learning_rate": 0.0007844600889446858, + "loss": 3.8679513931274414, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.007813206873834133, + "learning_rate": 0.0007832998978007873, + "loss": 3.860609531402588, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.007759213913232088, + "learning_rate": 0.0007821402961744448, + "loss": 3.803863048553467, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.008045864291489124, + "learning_rate": 0.0007809812849999293, + "loss": 3.858161449432373, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.008212368935346603, + "learning_rate": 0.0007798228652110347, + "loss": 3.8311336040496826, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.007982932962477207, + "learning_rate": 0.0007786650377410774, + "loss": 3.870339870452881, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.0071402136236429214, + "learning_rate": 0.0007775078035229002, + "loss": 3.7922263145446777, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.008564750663936138, + "learning_rate": 0.0007763511634888656, + "loss": 3.83981990814209, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.008078585378825665, + "learning_rate": 0.0007751951185708562, + "loss": 3.852938652038574, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.00801754929125309, + "learning_rate": 0.0007740396697002774, + "loss": 3.8434009552001953, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.008973612450063229, + "learning_rate": 0.0007728848178080535, + "loss": 3.8238563537597656, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.010199865326285362, + "learning_rate": 0.0007717305638246281, + "loss": 3.8891735076904297, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.009620471857488155, + "learning_rate": 0.0007705769086799617, + "loss": 3.8611769676208496, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.007981563918292522, + "learning_rate": 0.0007694238533035337, + "loss": 3.8496928215026855, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.008856567554175854, + "learning_rate": 0.0007682713986243415, + "loss": 3.8366098403930664, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.008323253132402897, + "learning_rate": 0.0007671195455708952, + "loss": 3.8224244117736816, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.00814893003553152, + "learning_rate": 0.0007659682950712228, + "loss": 3.8611154556274414, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.009146672673523426, + "learning_rate": 0.0007648176480528659, + "loss": 3.8258843421936035, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.007573423441499472, + "learning_rate": 0.0007636676054428805, + "loss": 3.888007640838623, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.00813984964042902, + "learning_rate": 0.0007625181681678343, + "loss": 3.860306739807129, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.007936778478324413, + "learning_rate": 0.0007613693371538078, + "loss": 3.8808555603027344, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.008434669114649296, + "learning_rate": 0.0007602211133263956, + "loss": 3.854154348373413, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.007857564836740494, + "learning_rate": 0.0007590734976106985, + "loss": 3.8240737915039062, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.007989327423274517, + "learning_rate": 0.0007579264909313315, + "loss": 3.850971221923828, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.008888393640518188, + "learning_rate": 0.0007567800942124152, + "loss": 3.862267255783081, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.38022518157959, + "eval_runtime": 51.4686, + "eval_samples_per_second": 47.446, + "eval_steps_per_second": 1.496, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.00952722318470478, + "learning_rate": 0.0007556343083775832, + "loss": 3.7705495357513428, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01099323108792305, + "learning_rate": 0.0007544891343499723, + "loss": 3.80812406539917, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.010298571549355984, + "learning_rate": 0.0007533445730522295, + "loss": 3.782794237136841, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.008139589801430702, + "learning_rate": 0.000752200625406507, + "loss": 3.7877392768859863, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.009031252935528755, + "learning_rate": 0.0007510572923344628, + "loss": 3.8181564807891846, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.00942459236830473, + "learning_rate": 0.0007499145747572603, + "loss": 3.807425022125244, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.00922947097569704, + "learning_rate": 0.0007487724735955648, + "loss": 3.7923264503479004, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.009551790542900562, + "learning_rate": 0.0007476309897695485, + "loss": 3.790083885192871, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.008996761403977871, + "learning_rate": 0.0007464901241988831, + "loss": 3.836491107940674, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.009766182862222195, + "learning_rate": 0.0007453498778027433, + "loss": 3.7829647064208984, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.007708575110882521, + "learning_rate": 0.0007442102514998058, + "loss": 3.803400993347168, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.008606902323663235, + "learning_rate": 0.0007430712462082464, + "loss": 3.810825824737549, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.009462927468121052, + "learning_rate": 0.0007419328628457416, + "loss": 3.784748077392578, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.010283366777002811, + "learning_rate": 0.0007407951023294655, + "loss": 3.744312286376953, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.008952794596552849, + "learning_rate": 0.0007396579655760917, + "loss": 3.8098015785217285, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.008446207270026207, + "learning_rate": 0.0007385214535017907, + "loss": 3.8306760787963867, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.007988285273313522, + "learning_rate": 0.0007373855670222302, + "loss": 3.7976696491241455, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.007586437277495861, + "learning_rate": 0.0007362503070525718, + "loss": 3.785891056060791, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.00775605533272028, + "learning_rate": 0.0007351156745074757, + "loss": 3.8176379203796387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.007727691903710365, + "learning_rate": 0.0007339816703010948, + "loss": 3.791743278503418, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.008311517536640167, + "learning_rate": 0.0007328482953470747, + "loss": 3.818769693374634, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.007610874716192484, + "learning_rate": 0.0007317155505585556, + "loss": 3.76772141456604, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.0078107016161084175, + "learning_rate": 0.0007305834368481699, + "loss": 3.822758197784424, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.006906864698976278, + "learning_rate": 0.0007294519551280412, + "loss": 3.832834005355835, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.008230660110712051, + "learning_rate": 0.0007283211063097829, + "loss": 3.85087513923645, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.008201118558645248, + "learning_rate": 0.0007271908913044997, + "loss": 3.7777247428894043, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.0070146857760846615, + "learning_rate": 0.000726061311022787, + "loss": 3.837517738342285, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.0076036700047552586, + "learning_rate": 0.0007249323663747254, + "loss": 3.8358592987060547, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.007238952908664942, + "learning_rate": 0.0007238040582698858, + "loss": 3.828089952468872, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.00825269054621458, + "learning_rate": 0.0007226763876173259, + "loss": 3.806016445159912, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.008368568494915962, + "learning_rate": 0.0007215493553255898, + "loss": 3.8163394927978516, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.00728471577167511, + "learning_rate": 0.0007204229623027063, + "loss": 3.835597038269043, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.00784830842167139, + "learning_rate": 0.0007192972094561897, + "loss": 3.8050971031188965, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.007714141625910997, + "learning_rate": 0.0007181720976930404, + "loss": 3.8031606674194336, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.009128103032708168, + "learning_rate": 0.000717047627919739, + "loss": 3.77107572555542, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.008033272810280323, + "learning_rate": 0.000715923801042251, + "loss": 3.8615059852600098, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.007732565049082041, + "learning_rate": 0.0007148006179660235, + "loss": 3.8277132511138916, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.007382966112345457, + "learning_rate": 0.0007136780795959853, + "loss": 3.803431987762451, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.007517020218074322, + "learning_rate": 0.0007125561868365442, + "loss": 3.799734354019165, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.008041434921324253, + "learning_rate": 0.0007114349405915895, + "loss": 3.7848258018493652, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.007875649258494377, + "learning_rate": 0.0007103143417644891, + "loss": 3.8136563301086426, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.008213640190660954, + "learning_rate": 0.000709194391258089, + "loss": 3.8518495559692383, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.00801054760813713, + "learning_rate": 0.0007080750899747139, + "loss": 3.7956271171569824, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.008064007386565208, + "learning_rate": 0.0007069564388161631, + "loss": 3.854595184326172, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.008300521411001682, + "learning_rate": 0.0007058384386837154, + "loss": 3.8203625679016113, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.008640496991574764, + "learning_rate": 0.000704721090478122, + "loss": 3.793574810028076, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.007684200536459684, + "learning_rate": 0.0007036043950996105, + "loss": 3.8564438819885254, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.008431905880570412, + "learning_rate": 0.0007024883534478823, + "loss": 3.8271119594573975, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.008137186989188194, + "learning_rate": 0.0007013729664221122, + "loss": 3.8219337463378906, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.0076744635589420795, + "learning_rate": 0.0007002582349209474, + "loss": 3.8353049755096436, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.008388740941882133, + "learning_rate": 0.0006991441598425065, + "loss": 3.795104503631592, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.007856911979615688, + "learning_rate": 0.0006980307420843798, + "loss": 3.818411111831665, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.008048887364566326, + "learning_rate": 0.000696917982543628, + "loss": 3.86599063873291, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.007949979975819588, + "learning_rate": 0.0006958058821167812, + "loss": 3.8026742935180664, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.008162821643054485, + "learning_rate": 0.0006946944416998389, + "loss": 3.833751678466797, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.008190765976905823, + "learning_rate": 0.0006935836621882682, + "loss": 3.8164236545562744, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.009714310988783836, + "learning_rate": 0.000692473544477005, + "loss": 3.8367366790771484, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.008075987920165062, + "learning_rate": 0.0006913640894604498, + "loss": 3.8108773231506348, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.009075167588889599, + "learning_rate": 0.0006902552980324713, + "loss": 3.8444485664367676, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.008678440935909748, + "learning_rate": 0.0006891471710864022, + "loss": 3.8384461402893066, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.007705411408096552, + "learning_rate": 0.0006880397095150414, + "loss": 3.825443744659424, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.00782099086791277, + "learning_rate": 0.0006869329142106488, + "loss": 3.8331687450408936, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.007542501203715801, + "learning_rate": 0.0006858267860649509, + "loss": 3.7905991077423096, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.008315361104905605, + "learning_rate": 0.0006847213259691356, + "loss": 3.861527442932129, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.008529976941645145, + "learning_rate": 0.0006836165348138508, + "loss": 3.8332972526550293, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.008117211051285267, + "learning_rate": 0.0006825124134892072, + "loss": 3.7961089611053467, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.008409497328102589, + "learning_rate": 0.0006814089628847755, + "loss": 3.812218189239502, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.007592010777443647, + "learning_rate": 0.0006803061838895864, + "loss": 3.801304340362549, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.007820053026080132, + "learning_rate": 0.0006792040773921283, + "loss": 3.8517117500305176, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.0073044621385633945, + "learning_rate": 0.0006781026442803479, + "loss": 3.8253560066223145, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.007546546868979931, + "learning_rate": 0.0006770018854416522, + "loss": 3.800802230834961, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.008878474123775959, + "learning_rate": 0.0006759018017629007, + "loss": 3.8133859634399414, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.375829219818115, + "eval_runtime": 51.3756, + "eval_samples_per_second": 47.532, + "eval_steps_per_second": 1.499, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.00814585480839014, + "learning_rate": 0.0006748023941304115, + "loss": 3.7970919609069824, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.00824262760579586, + "learning_rate": 0.0006737036634299577, + "loss": 3.750131607055664, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.008675593882799149, + "learning_rate": 0.000672605610546767, + "loss": 3.791456699371338, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.009052547626197338, + "learning_rate": 0.00067150823636552, + "loss": 3.7762060165405273, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.00818016566336155, + "learning_rate": 0.0006704115417703514, + "loss": 3.750631332397461, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.00800115056335926, + "learning_rate": 0.0006693155276448489, + "loss": 3.811519145965576, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.009032433852553368, + "learning_rate": 0.0006682201948720509, + "loss": 3.7805609703063965, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.0077886078506708145, + "learning_rate": 0.0006671255443344474, + "loss": 3.749366521835327, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.008603204973042011, + "learning_rate": 0.0006660315769139774, + "loss": 3.7686312198638916, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.008361323736608028, + "learning_rate": 0.0006649382934920329, + "loss": 3.740269660949707, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.00948982872068882, + "learning_rate": 0.0006638456949494507, + "loss": 3.7874045372009277, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.008958354592323303, + "learning_rate": 0.0006627537821665185, + "loss": 3.78987455368042, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.008702470920979977, + "learning_rate": 0.0006616625560229704, + "loss": 3.751061201095581, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.008424391970038414, + "learning_rate": 0.000660572017397988, + "loss": 3.80816650390625, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.009022071026265621, + "learning_rate": 0.0006594821671701989, + "loss": 3.726067304611206, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.007974586449563503, + "learning_rate": 0.0006583930062176739, + "loss": 3.78564453125, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.007887622341513634, + "learning_rate": 0.0006573045354179329, + "loss": 3.7993059158325195, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.007782666943967342, + "learning_rate": 0.0006562167556479352, + "loss": 3.7440507411956787, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.007923880591988564, + "learning_rate": 0.000655129667784086, + "loss": 3.7920784950256348, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.007850701920688152, + "learning_rate": 0.0006540432727022325, + "loss": 3.7833633422851562, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.007855751551687717, + "learning_rate": 0.0006529575712776633, + "loss": 3.812349319458008, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.007832377217710018, + "learning_rate": 0.0006518725643851091, + "loss": 3.7815146446228027, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.006997724529355764, + "learning_rate": 0.0006507882528987395, + "loss": 3.795346975326538, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.007433670572936535, + "learning_rate": 0.0006497046376921648, + "loss": 3.841017246246338, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.0075767990201711655, + "learning_rate": 0.0006486217196384344, + "loss": 3.7913427352905273, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.007551983464509249, + "learning_rate": 0.0006475394996100359, + "loss": 3.765105962753296, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.007235546596348286, + "learning_rate": 0.0006464579784788943, + "loss": 3.7668676376342773, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.007872921414673328, + "learning_rate": 0.0006453771571163716, + "loss": 3.770536422729492, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.007598782889544964, + "learning_rate": 0.0006442970363932667, + "loss": 3.796818733215332, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.007902889512479305, + "learning_rate": 0.0006432176171798124, + "loss": 3.824531078338623, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.008964902721345425, + "learning_rate": 0.0006421389003456776, + "loss": 3.781778335571289, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.009936625137925148, + "learning_rate": 0.000641060886759965, + "loss": 3.750248670578003, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.008381214924156666, + "learning_rate": 0.0006399835772912106, + "loss": 3.797954797744751, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.007752922363579273, + "learning_rate": 0.000638906972807384, + "loss": 3.789350748062134, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.009501751512289047, + "learning_rate": 0.0006378310741758841, + "loss": 3.791079521179199, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.00936005637049675, + "learning_rate": 0.0006367558822635453, + "loss": 3.78501033782959, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.008370982483029366, + "learning_rate": 0.0006356813979366284, + "loss": 3.792647361755371, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.00782831758260727, + "learning_rate": 0.000634607622060827, + "loss": 3.772569179534912, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.00830658059567213, + "learning_rate": 0.0006335345555012627, + "loss": 3.7787656784057617, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.009015833027660847, + "learning_rate": 0.000632462199122486, + "loss": 3.787418842315674, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.008993860334157944, + "learning_rate": 0.0006313905537884752, + "loss": 3.818718433380127, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.007954884320497513, + "learning_rate": 0.0006303196203626344, + "loss": 3.8100008964538574, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.007742907386273146, + "learning_rate": 0.0006292493997077973, + "loss": 3.7786598205566406, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.008748230524361134, + "learning_rate": 0.0006281798926862195, + "loss": 3.798635482788086, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.008306438103318214, + "learning_rate": 0.000627111100159585, + "loss": 3.827213764190674, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.008027350530028343, + "learning_rate": 0.0006260430229889987, + "loss": 3.7682337760925293, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.007779776118695736, + "learning_rate": 0.0006249756620349931, + "loss": 3.7515475749969482, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.008748390711843967, + "learning_rate": 0.000623909018157521, + "loss": 3.7963085174560547, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.008587505668401718, + "learning_rate": 0.0006228430922159577, + "loss": 3.7767529487609863, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.007172323763370514, + "learning_rate": 0.0006217778850691007, + "loss": 3.79768705368042, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.007224083412438631, + "learning_rate": 0.000620713397575168, + "loss": 3.785390853881836, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.008209837600588799, + "learning_rate": 0.0006196496305917987, + "loss": 3.809828281402588, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.007760563865303993, + "learning_rate": 0.0006185865849760491, + "loss": 3.838716506958008, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.007362260017544031, + "learning_rate": 0.0006175242615843977, + "loss": 3.7840137481689453, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.008719335310161114, + "learning_rate": 0.0006164626612727381, + "loss": 3.8228893280029297, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.00856029987335205, + "learning_rate": 0.0006154017848963826, + "loss": 3.80448317527771, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.008078506216406822, + "learning_rate": 0.0006143416333100606, + "loss": 3.7881453037261963, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.007957885973155499, + "learning_rate": 0.0006132822073679164, + "loss": 3.814870834350586, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.009362339973449707, + "learning_rate": 0.0006122235079235114, + "loss": 3.798335075378418, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.008272632025182247, + "learning_rate": 0.0006111655358298194, + "loss": 3.8212976455688477, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.007548908703029156, + "learning_rate": 0.0006101082919392301, + "loss": 3.8312830924987793, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.007543765474110842, + "learning_rate": 0.0006090517771035457, + "loss": 3.8302183151245117, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.007520351093262434, + "learning_rate": 0.000607995992173981, + "loss": 3.7898802757263184, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.007168850861489773, + "learning_rate": 0.000606940938001163, + "loss": 3.8078577518463135, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.007420001085847616, + "learning_rate": 0.0006058866154351296, + "loss": 3.799571990966797, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.007621012162417173, + "learning_rate": 0.0006048330253253303, + "loss": 3.833843946456909, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.007661828305572271, + "learning_rate": 0.0006037801685206222, + "loss": 3.820556640625, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.007657836657017469, + "learning_rate": 0.0006027280458692738, + "loss": 3.798994541168213, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.007722760550677776, + "learning_rate": 0.0006016766582189613, + "loss": 3.8589251041412354, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.00789247453212738, + "learning_rate": 0.0006006260064167685, + "loss": 3.8673787117004395, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.009137817658483982, + "learning_rate": 0.0005995760913091874, + "loss": 3.801938056945801, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.008374894969165325, + "learning_rate": 0.0005985269137421141, + "loss": 3.820045232772827, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.3730878829956055, + "eval_runtime": 48.1607, + "eval_samples_per_second": 50.705, + "eval_steps_per_second": 1.599, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.009492472745478153, + "learning_rate": 0.0005974784745608536, + "loss": 3.738617420196533, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.010017318651080132, + "learning_rate": 0.0005964307746101135, + "loss": 3.7551231384277344, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.008958659134805202, + "learning_rate": 0.000595383814734007, + "loss": 3.7739763259887695, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.008034365251660347, + "learning_rate": 0.0005943375957760505, + "loss": 3.7678627967834473, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.008489780128002167, + "learning_rate": 0.000593292118579164, + "loss": 3.720778465270996, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.009107839316129684, + "learning_rate": 0.0005922473839856699, + "loss": 3.7224440574645996, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.009295578114688396, + "learning_rate": 0.0005912033928372906, + "loss": 3.7695257663726807, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.007896862924098969, + "learning_rate": 0.0005901601459751532, + "loss": 3.742567300796509, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.008227395825088024, + "learning_rate": 0.0005891176442397808, + "loss": 3.7547202110290527, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.009326796047389507, + "learning_rate": 0.0005880758884710992, + "loss": 3.724808931350708, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.007786506786942482, + "learning_rate": 0.0005870348795084319, + "loss": 3.7403640747070312, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.00821041688323021, + "learning_rate": 0.0005859946181905009, + "loss": 3.770012617111206, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.009544595144689083, + "learning_rate": 0.000584955105355427, + "loss": 3.7378525733947754, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.008973932825028896, + "learning_rate": 0.0005839163418407254, + "loss": 3.7184295654296875, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.009023688733577728, + "learning_rate": 0.00058287832848331, + "loss": 3.7388229370117188, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.008151066489517689, + "learning_rate": 0.0005818410661194889, + "loss": 3.7632875442504883, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.008495030924677849, + "learning_rate": 0.0005808045555849668, + "loss": 3.8114736080169678, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.008739210665225983, + "learning_rate": 0.0005797687977148394, + "loss": 3.7647829055786133, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.008062886074185371, + "learning_rate": 0.0005787337933435999, + "loss": 3.782172441482544, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.007615398149937391, + "learning_rate": 0.0005776995433051326, + "loss": 3.7564642429351807, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.007938290014863014, + "learning_rate": 0.0005766660484327133, + "loss": 3.756208896636963, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.008517666719853878, + "learning_rate": 0.0005756333095590102, + "loss": 3.76729416847229, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.007426336407661438, + "learning_rate": 0.0005746013275160827, + "loss": 3.74001407623291, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.007664753124117851, + "learning_rate": 0.0005735701031353804, + "loss": 3.7819876670837402, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.007903067395091057, + "learning_rate": 0.0005725396372477414, + "loss": 3.781219959259033, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.007251975126564503, + "learning_rate": 0.0005715099306833931, + "loss": 3.763620615005493, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.007422176189720631, + "learning_rate": 0.0005704809842719533, + "loss": 3.775099754333496, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.009161600843071938, + "learning_rate": 0.0005694527988424239, + "loss": 3.792268991470337, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.008330529555678368, + "learning_rate": 0.0005684253752231958, + "loss": 3.7967824935913086, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.0073860278353095055, + "learning_rate": 0.0005673987142420458, + "loss": 3.76505970954895, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.008462096564471722, + "learning_rate": 0.0005663728167261363, + "loss": 3.740450143814087, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.00838128849864006, + "learning_rate": 0.0005653476835020138, + "loss": 3.7863235473632812, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.007381014991551638, + "learning_rate": 0.0005643233153956092, + "loss": 3.772179126739502, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.007363761775195599, + "learning_rate": 0.0005632997132322392, + "loss": 3.7793867588043213, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.008684969507157803, + "learning_rate": 0.0005622768778366001, + "loss": 3.7736339569091797, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.007934059016406536, + "learning_rate": 0.0005612548100327722, + "loss": 3.8040053844451904, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.007136524189263582, + "learning_rate": 0.0005602335106442172, + "loss": 3.7589468955993652, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.00789916142821312, + "learning_rate": 0.0005592129804937779, + "loss": 3.7953195571899414, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.007919179275631905, + "learning_rate": 0.0005581932204036761, + "loss": 3.7539429664611816, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.007734223268926144, + "learning_rate": 0.0005571742311955147, + "loss": 3.7620325088500977, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.007005508057773113, + "learning_rate": 0.0005561560136902749, + "loss": 3.762538433074951, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.008137146942317486, + "learning_rate": 0.0005551385687083161, + "loss": 3.773789405822754, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.007310070563107729, + "learning_rate": 0.0005541218970693757, + "loss": 3.8077168464660645, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.007335536181926727, + "learning_rate": 0.0005531059995925661, + "loss": 3.7761921882629395, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.008022893220186234, + "learning_rate": 0.00055209087709638, + "loss": 3.7769880294799805, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.008027128875255585, + "learning_rate": 0.0005510765303986814, + "loss": 3.7709951400756836, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.00697955721989274, + "learning_rate": 0.0005500629603167118, + "loss": 3.7855210304260254, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.007455077022314072, + "learning_rate": 0.0005490501676670861, + "loss": 3.7850632667541504, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.0069657545536756516, + "learning_rate": 0.0005480381532657932, + "loss": 3.742671251296997, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.007609451189637184, + "learning_rate": 0.0005470269179281953, + "loss": 3.793771266937256, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.007437967229634523, + "learning_rate": 0.0005460164624690255, + "loss": 3.778846263885498, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.007123028859496117, + "learning_rate": 0.00054500678770239, + "loss": 3.768375873565674, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.008176218718290329, + "learning_rate": 0.0005439978944417654, + "loss": 3.82069730758667, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.006943001411855221, + "learning_rate": 0.000542989783499999, + "loss": 3.7731552124023438, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.0072205523028969765, + "learning_rate": 0.0005419824556893069, + "loss": 3.7724058628082275, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.007656400091946125, + "learning_rate": 0.0005409759118212758, + "loss": 3.780989646911621, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.006972154602408409, + "learning_rate": 0.0005399701527068599, + "loss": 3.7971787452697754, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.007674331311136484, + "learning_rate": 0.0005389651791563806, + "loss": 3.798673391342163, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.006952045019716024, + "learning_rate": 0.0005379609919795268, + "loss": 3.7982821464538574, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.007882075384259224, + "learning_rate": 0.0005369575919853545, + "loss": 3.825962781906128, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.007143386639654636, + "learning_rate": 0.0005359549799822854, + "loss": 3.826598644256592, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.007587207015603781, + "learning_rate": 0.0005349531567781041, + "loss": 3.80623197555542, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.007677781395614147, + "learning_rate": 0.0005339521231799632, + "loss": 3.7987256050109863, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.007372343447059393, + "learning_rate": 0.0005329518799943773, + "loss": 3.7733969688415527, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.00861304346472025, + "learning_rate": 0.0005319524280272231, + "loss": 3.7903685569763184, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.00816237274557352, + "learning_rate": 0.0005309537680837418, + "loss": 3.77058744430542, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.007624274119734764, + "learning_rate": 0.0005299559009685356, + "loss": 3.787982940673828, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.007416521664708853, + "learning_rate": 0.0005289588274855683, + "loss": 3.7991762161254883, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.007461614441126585, + "learning_rate": 0.0005279625484381632, + "loss": 3.8036863803863525, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.007766335736960173, + "learning_rate": 0.0005269670646290043, + "loss": 3.809243679046631, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.0073394919745624065, + "learning_rate": 0.0005259723768601365, + "loss": 3.756606101989746, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.008771199733018875, + "learning_rate": 0.0005249784859329602, + "loss": 3.773951768875122, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.370637893676758, + "eval_runtime": 49.4904, + "eval_samples_per_second": 49.343, + "eval_steps_per_second": 1.556, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.008164486847817898, + "learning_rate": 0.0005239853926482358, + "loss": 3.7619595527648926, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.007980132475495338, + "learning_rate": 0.0005229930978060806, + "loss": 3.7398204803466797, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.0077056242153048515, + "learning_rate": 0.0005220016022059692, + "loss": 3.699580669403076, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.0074698119424283504, + "learning_rate": 0.0005210109066467308, + "loss": 3.7299928665161133, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.008384468965232372, + "learning_rate": 0.0005200210119265512, + "loss": 3.704989433288574, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.008066141977906227, + "learning_rate": 0.0005190319188429709, + "loss": 3.751697540283203, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.007213297300040722, + "learning_rate": 0.0005180436281928841, + "loss": 3.740692377090454, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.007343464996665716, + "learning_rate": 0.0005170561407725396, + "loss": 3.7060189247131348, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.00783725269138813, + "learning_rate": 0.0005160694573775363, + "loss": 3.7414426803588867, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.007640232797712088, + "learning_rate": 0.0005150835788028294, + "loss": 3.7451486587524414, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.008129071444272995, + "learning_rate": 0.0005140985058427218, + "loss": 3.734839916229248, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.006920623127371073, + "learning_rate": 0.0005131142392908697, + "loss": 3.7138614654541016, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.00798740889877081, + "learning_rate": 0.0005121307799402787, + "loss": 3.76496958732605, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.006689135450869799, + "learning_rate": 0.0005111481285833045, + "loss": 3.7707948684692383, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.007544311694800854, + "learning_rate": 0.0005101662860116516, + "loss": 3.7393479347229004, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.007487910334020853, + "learning_rate": 0.0005091852530163717, + "loss": 3.7486305236816406, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.00772349676117301, + "learning_rate": 0.0005082050303878673, + "loss": 3.753629207611084, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.007883663289248943, + "learning_rate": 0.0005072256189158844, + "loss": 3.7497611045837402, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.008153906092047691, + "learning_rate": 0.0005062470193895178, + "loss": 3.7459521293640137, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.007083641365170479, + "learning_rate": 0.0005052692325972075, + "loss": 3.725599765777588, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.007102683652192354, + "learning_rate": 0.0005042922593267385, + "loss": 3.7724428176879883, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.008104706183075905, + "learning_rate": 0.0005033161003652409, + "loss": 3.7386724948883057, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.0073765418492257595, + "learning_rate": 0.0005023407564991878, + "loss": 3.7478132247924805, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.007110628765076399, + "learning_rate": 0.0005013662285143963, + "loss": 3.771925687789917, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.007323848083615303, + "learning_rate": 0.000500392517196026, + "loss": 3.7460904121398926, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.007830807007849216, + "learning_rate": 0.0004994196233285786, + "loss": 3.751222610473633, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.007710007485002279, + "learning_rate": 0.0004984475476958973, + "loss": 3.739858627319336, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.007293973118066788, + "learning_rate": 0.0004974762910811655, + "loss": 3.7410120964050293, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.0073824538849294186, + "learning_rate": 0.0004965058542669078, + "loss": 3.711397647857666, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.007156423293054104, + "learning_rate": 0.0004955362380349866, + "loss": 3.7881593704223633, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.007705818396061659, + "learning_rate": 0.0004945674431666047, + "loss": 3.7639269828796387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.0075300210155546665, + "learning_rate": 0.0004935994704423026, + "loss": 3.7562999725341797, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.007137116510421038, + "learning_rate": 0.000492632320641959, + "loss": 3.7659149169921875, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.007225630804896355, + "learning_rate": 0.0004916659945447878, + "loss": 3.7625155448913574, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.008027992211282253, + "learning_rate": 0.0004907004929293408, + "loss": 3.746478796005249, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.007582477759569883, + "learning_rate": 0.0004897358165735066, + "loss": 3.759408950805664, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.007575645577162504, + "learning_rate": 0.000488771966254506, + "loss": 3.7430644035339355, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.008404700085520744, + "learning_rate": 0.0004878089427488963, + "loss": 3.7727198600769043, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.007118664216250181, + "learning_rate": 0.00048684674683256815, + "loss": 3.747460126876831, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.007220778148621321, + "learning_rate": 0.00048588537928074584, + "loss": 3.746413230895996, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.0076194945722818375, + "learning_rate": 0.0004849248408679849, + "loss": 3.7808125019073486, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.007285538129508495, + "learning_rate": 0.00048396513236817387, + "loss": 3.7750301361083984, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.007698678877204657, + "learning_rate": 0.00048300625455453383, + "loss": 3.7474594116210938, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.007687424309551716, + "learning_rate": 0.0004820482081996141, + "loss": 3.7929720878601074, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.007363605313003063, + "learning_rate": 0.0004810909940752959, + "loss": 3.75911021232605, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.007791243493556976, + "learning_rate": 0.00048013461295278836, + "loss": 3.7567291259765625, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.007375436834990978, + "learning_rate": 0.0004791790656026323, + "loss": 3.7784414291381836, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.007330593653023243, + "learning_rate": 0.0004782243527946934, + "loss": 3.7533810138702393, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.006926248781383038, + "learning_rate": 0.00047727047529816694, + "loss": 3.7795910835266113, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.007102136500179768, + "learning_rate": 0.00047631743388157496, + "loss": 3.812549114227295, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.006676106248050928, + "learning_rate": 0.00047536522931276555, + "loss": 3.7578792572021484, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.006926767062395811, + "learning_rate": 0.0004744138623589134, + "loss": 3.782416820526123, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.006894777063280344, + "learning_rate": 0.000473463333786516, + "loss": 3.7840042114257812, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.007393381092697382, + "learning_rate": 0.00047251364436139923, + "loss": 3.7532780170440674, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.006510896608233452, + "learning_rate": 0.0004715647948487092, + "loss": 3.7618460655212402, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.007283828221261501, + "learning_rate": 0.0004706167860129171, + "loss": 3.768498420715332, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.00655644154176116, + "learning_rate": 0.0004696696186178168, + "loss": 3.7724053859710693, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.007338619790971279, + "learning_rate": 0.0004687232934265239, + "loss": 3.7814595699310303, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.006839970126748085, + "learning_rate": 0.0004677778112014758, + "loss": 3.771808624267578, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.006776005961000919, + "learning_rate": 0.0004668331727044297, + "loss": 3.786924123764038, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.007011357229202986, + "learning_rate": 0.00046588937869646415, + "loss": 3.787672996520996, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.006482086144387722, + "learning_rate": 0.0004649464299379771, + "loss": 3.7794556617736816, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.0069654518738389015, + "learning_rate": 0.000464004327188685, + "loss": 3.7520360946655273, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.006526473443955183, + "learning_rate": 0.0004630630712076235, + "loss": 3.734144926071167, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.006822322495281696, + "learning_rate": 0.00046212266275314504, + "loss": 3.7614707946777344, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.006827171426266432, + "learning_rate": 0.00046118310258292045, + "loss": 3.755218029022217, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.00706718722358346, + "learning_rate": 0.00046024439145393546, + "loss": 3.7397823333740234, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.00689381267875433, + "learning_rate": 0.00045930653012249314, + "loss": 3.758922576904297, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.007069882471114397, + "learning_rate": 0.0004583695193442113, + "loss": 3.759767532348633, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.006889182608574629, + "learning_rate": 0.0004574333598740228, + "loss": 3.7555460929870605, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.006866258103400469, + "learning_rate": 0.00045649805246617475, + "loss": 3.762068271636963, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.008696220815181732, + "learning_rate": 0.000455563597874226, + "loss": 3.7110466957092285, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.369833469390869, + "eval_runtime": 48.7852, + "eval_samples_per_second": 50.056, + "eval_steps_per_second": 1.578, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.0077184769324958324, + "learning_rate": 0.0004546299968510516, + "loss": 3.704669713973999, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.00846019759774208, + "learning_rate": 0.0004536972501488358, + "loss": 3.701132297515869, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.008113945834338665, + "learning_rate": 0.00045276535851907613, + "loss": 3.699582099914551, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.0072370911948382854, + "learning_rate": 0.0004518343227125807, + "loss": 3.698126792907715, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.00817187875509262, + "learning_rate": 0.0004509041434794685, + "loss": 3.738393783569336, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.007495569530874491, + "learning_rate": 0.00044997482156916845, + "loss": 3.727083683013916, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.007855897769331932, + "learning_rate": 0.0004490463577304171, + "loss": 3.716137170791626, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.007817867211997509, + "learning_rate": 0.0004481187527112628, + "loss": 3.750791072845459, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.007712380029261112, + "learning_rate": 0.00044719200725905887, + "loss": 3.720554828643799, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.00859921332448721, + "learning_rate": 0.00044626612212046816, + "loss": 3.742496967315674, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.007692081853747368, + "learning_rate": 0.0004453410980414582, + "loss": 3.703803300857544, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.007734883576631546, + "learning_rate": 0.00044441693576730565, + "loss": 3.7371301651000977, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.009022004902362823, + "learning_rate": 0.0004434936360425916, + "loss": 3.727294683456421, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.008058780804276466, + "learning_rate": 0.00044257119961120067, + "loss": 3.7583885192871094, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.00698013324290514, + "learning_rate": 0.00044164962721632376, + "loss": 3.7312440872192383, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.007999536581337452, + "learning_rate": 0.00044072891960045523, + "loss": 3.6972708702087402, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.008125107735395432, + "learning_rate": 0.0004398090775053929, + "loss": 3.7123398780822754, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.006736223120242357, + "learning_rate": 0.0004388901016722357, + "loss": 3.749678134918213, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.009154478088021278, + "learning_rate": 0.000437971992841387, + "loss": 3.722728967666626, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.008070871233940125, + "learning_rate": 0.00043705475175255087, + "loss": 3.7297017574310303, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.007498566992580891, + "learning_rate": 0.0004361383791447307, + "loss": 3.7286131381988525, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.007642887998372316, + "learning_rate": 0.00043522287575623183, + "loss": 3.750898838043213, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.007800443097949028, + "learning_rate": 0.0004343082423246588, + "loss": 3.719512462615967, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.007393732666969299, + "learning_rate": 0.0004333944795869159, + "loss": 3.7136006355285645, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.008559555746614933, + "learning_rate": 0.0004324815882792042, + "loss": 3.7087020874023438, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.007379850372672081, + "learning_rate": 0.0004315695691370241, + "loss": 3.7638165950775146, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.007374506443738937, + "learning_rate": 0.00043065842289517303, + "loss": 3.7208821773529053, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.00755207147449255, + "learning_rate": 0.0004297481502877451, + "loss": 3.738490581512451, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.007752610836178064, + "learning_rate": 0.0004288387520481306, + "loss": 3.7401833534240723, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.006907653529196978, + "learning_rate": 0.0004279302289090151, + "loss": 3.7754998207092285, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.007771312724798918, + "learning_rate": 0.0004270225816023797, + "loss": 3.73626446723938, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.007395982276648283, + "learning_rate": 0.0004261158108594987, + "loss": 3.7159175872802734, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.007196040358394384, + "learning_rate": 0.0004252099174109411, + "loss": 3.750246047973633, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.008241173811256886, + "learning_rate": 0.0004243049019865692, + "loss": 3.70859694480896, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.0072947596199810505, + "learning_rate": 0.00042340076531553755, + "loss": 3.7385268211364746, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.00782464537769556, + "learning_rate": 0.00042249750812629297, + "loss": 3.75885272026062, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.007813232019543648, + "learning_rate": 0.0004215951311465722, + "loss": 3.71810245513916, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.006900500506162643, + "learning_rate": 0.00042069363510340604, + "loss": 3.7365174293518066, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.007812316995114088, + "learning_rate": 0.0004197930207231119, + "loss": 3.774794101715088, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.007018622010946274, + "learning_rate": 0.00041889328873129933, + "loss": 3.7112040519714355, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.006895892322063446, + "learning_rate": 0.0004179944398528659, + "loss": 3.7497243881225586, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.0071940235793590546, + "learning_rate": 0.0004170964748119979, + "loss": 3.7566123008728027, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.006776201073080301, + "learning_rate": 0.0004161993943321702, + "loss": 3.7683968544006348, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.006886148825287819, + "learning_rate": 0.0004153031991361428, + "loss": 3.759763240814209, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.006466342136263847, + "learning_rate": 0.0004144078899459659, + "loss": 3.736006498336792, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.007055724039673805, + "learning_rate": 0.0004135134674829725, + "loss": 3.70645809173584, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.006948428228497505, + "learning_rate": 0.00041261993246778303, + "loss": 3.7365002632141113, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.006491438951343298, + "learning_rate": 0.00041172728562030255, + "loss": 3.735001564025879, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.006851755082607269, + "learning_rate": 0.0004108355276597209, + "loss": 3.720893383026123, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.007266475819051266, + "learning_rate": 0.00040994465930451114, + "loss": 3.7746100425720215, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.007112737279385328, + "learning_rate": 0.00040905468127242926, + "loss": 3.7441720962524414, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.006514255423098803, + "learning_rate": 0.000408165594280515, + "loss": 3.7333898544311523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.007172788958996534, + "learning_rate": 0.0004072773990450896, + "loss": 3.7675576210021973, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.00684684282168746, + "learning_rate": 0.0004063900962817564, + "loss": 3.7141308784484863, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.006578145083039999, + "learning_rate": 0.00040550368670539777, + "loss": 3.7256317138671875, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.007047492545098066, + "learning_rate": 0.00040461817103017926, + "loss": 3.7618556022644043, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.007036720868200064, + "learning_rate": 0.00040373354996954487, + "loss": 3.7397944927215576, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.00745626725256443, + "learning_rate": 0.00040284982423621714, + "loss": 3.7831153869628906, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.007454817183315754, + "learning_rate": 0.00040196699454219797, + "loss": 3.7409911155700684, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.0075650583021342754, + "learning_rate": 0.00040108506159876756, + "loss": 3.7074005603790283, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.006908242590725422, + "learning_rate": 0.0004002040261164838, + "loss": 3.767744779586792, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.0072857895866036415, + "learning_rate": 0.0003993238888051801, + "loss": 3.7349488735198975, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.0072942771948874, + "learning_rate": 0.00039844465037396716, + "loss": 3.75968861579895, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.0071496241725981236, + "learning_rate": 0.00039756631153123296, + "loss": 3.8237380981445312, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.006803249940276146, + "learning_rate": 0.00039668887298463783, + "loss": 3.7468650341033936, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.007433352991938591, + "learning_rate": 0.00039581233544111867, + "loss": 3.783236503601074, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.006633558310568333, + "learning_rate": 0.0003949366996068858, + "loss": 3.75020694732666, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.007415154483169317, + "learning_rate": 0.00039406196618742376, + "loss": 3.7545979022979736, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.006997843272984028, + "learning_rate": 0.0003931881358874886, + "loss": 3.7404396533966064, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.007417330518364906, + "learning_rate": 0.0003923152094111095, + "loss": 3.7575364112854004, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.006936701480299234, + "learning_rate": 0.0003914431874615889, + "loss": 3.771845817565918, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.008293208666145802, + "learning_rate": 0.00039057207074149795, + "loss": 3.746669292449951, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.370476245880127, + "eval_runtime": 50.9368, + "eval_samples_per_second": 47.942, + "eval_steps_per_second": 1.512, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.0077544888481497765, + "learning_rate": 0.0003897018599526799, + "loss": 3.6915597915649414, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.008674043230712414, + "learning_rate": 0.000388832555796248, + "loss": 3.67698073387146, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.00724575063213706, + "learning_rate": 0.00038796415897258523, + "loss": 3.6838502883911133, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.00849415734410286, + "learning_rate": 0.00038709667018134246, + "loss": 3.69064998626709, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.00769044877961278, + "learning_rate": 0.0003862300901214403, + "loss": 3.670502185821533, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.007846124470233917, + "learning_rate": 0.00038536441949106647, + "loss": 3.767771005630493, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.007555077783763409, + "learning_rate": 0.000384499658987676, + "loss": 3.7001633644104004, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.007834251970052719, + "learning_rate": 0.0003836358093079915, + "loss": 3.704331874847412, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.007608097977936268, + "learning_rate": 0.00038277287114799974, + "loss": 3.7108519077301025, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.007420446258038282, + "learning_rate": 0.0003819108452029561, + "loss": 3.697843313217163, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.008091001771390438, + "learning_rate": 0.000381049732167378, + "loss": 3.7380661964416504, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.007265515625476837, + "learning_rate": 0.00038018953273504907, + "loss": 3.680748462677002, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.007735357619822025, + "learning_rate": 0.00037933024759901636, + "loss": 3.6648077964782715, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.008188895881175995, + "learning_rate": 0.0003784718774515907, + "loss": 3.706609010696411, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.00816027820110321, + "learning_rate": 0.00037761442298434554, + "loss": 3.7368297576904297, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.006931936834007502, + "learning_rate": 0.00037675788488811566, + "loss": 3.7121026515960693, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.007615855894982815, + "learning_rate": 0.00037590226385299897, + "loss": 3.7379372119903564, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.0073505607433617115, + "learning_rate": 0.0003750475605683535, + "loss": 3.7330894470214844, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.007151979021728039, + "learning_rate": 0.00037419377572279897, + "loss": 3.713761568069458, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.007310053799301386, + "learning_rate": 0.00037334091000421326, + "loss": 3.705451011657715, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.0072968751192092896, + "learning_rate": 0.0003724889640997357, + "loss": 3.7550907135009766, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.007197601720690727, + "learning_rate": 0.00037163793869576404, + "loss": 3.7346463203430176, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.007823911495506763, + "learning_rate": 0.00037078783447795307, + "loss": 3.7257699966430664, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.007118315435945988, + "learning_rate": 0.0003699386521312171, + "loss": 3.7014496326446533, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.0069230347871780396, + "learning_rate": 0.00036909039233972666, + "loss": 3.6709237098693848, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.007836844772100449, + "learning_rate": 0.00036824305578691, + "loss": 3.695725440979004, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.007123398594558239, + "learning_rate": 0.00036739664315544956, + "loss": 3.7278594970703125, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.007792143151164055, + "learning_rate": 0.00036655115512728586, + "loss": 3.7291927337646484, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.007245976477861404, + "learning_rate": 0.00036570659238361345, + "loss": 3.7183046340942383, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.007778054103255272, + "learning_rate": 0.0003648629556048804, + "loss": 3.7485761642456055, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.007539471611380577, + "learning_rate": 0.00036402024547079007, + "loss": 3.7273335456848145, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.007554007228463888, + "learning_rate": 0.0003631784626602987, + "loss": 3.684359073638916, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.007862667553126812, + "learning_rate": 0.00036233760785161606, + "loss": 3.7223072052001953, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.00869659148156643, + "learning_rate": 0.0003614976817222028, + "loss": 3.7475168704986572, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.00696196174249053, + "learning_rate": 0.00036065868494877226, + "loss": 3.7773914337158203, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.008952745236456394, + "learning_rate": 0.00035982061820729005, + "loss": 3.6761155128479004, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.008341601118445396, + "learning_rate": 0.00035898348217297034, + "loss": 3.746553421020508, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.007268146146088839, + "learning_rate": 0.00035814727752027885, + "loss": 3.6921701431274414, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.007401647046208382, + "learning_rate": 0.00035731200492293064, + "loss": 3.722433090209961, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.006969883106648922, + "learning_rate": 0.00035647766505389006, + "loss": 3.7449448108673096, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.007616426330059767, + "learning_rate": 0.00035564425858536895, + "loss": 3.725113868713379, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.006883042864501476, + "learning_rate": 0.00035481178618882814, + "loss": 3.717106342315674, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.007106961216777563, + "learning_rate": 0.00035398024853497574, + "loss": 3.74045467376709, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.007402827963232994, + "learning_rate": 0.00035314964629376664, + "loss": 3.7288095951080322, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.006770852487534285, + "learning_rate": 0.00035231998013440227, + "loss": 3.7663369178771973, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.007622968405485153, + "learning_rate": 0.0003514912507253287, + "loss": 3.747084140777588, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.006876434665173292, + "learning_rate": 0.00035066345873423975, + "loss": 3.7591025829315186, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.007071215659379959, + "learning_rate": 0.00034983660482807116, + "loss": 3.7583885192871094, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.007130554411560297, + "learning_rate": 0.00034901068967300465, + "loss": 3.6950902938842773, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.006981188431382179, + "learning_rate": 0.00034818571393446524, + "loss": 3.7390527725219727, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.0067793820053339005, + "learning_rate": 0.0003473616782771209, + "loss": 3.703673839569092, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.00674026133492589, + "learning_rate": 0.0003465385833648827, + "loss": 3.6988182067871094, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.006553220096975565, + "learning_rate": 0.0003457164298609021, + "loss": 3.7367048263549805, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.007079749833792448, + "learning_rate": 0.000344895218427575, + "loss": 3.773226022720337, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.006091540213674307, + "learning_rate": 0.00034407494972653547, + "loss": 3.747929096221924, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.006783034652471542, + "learning_rate": 0.00034325562441865935, + "loss": 3.7154481410980225, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.006683558225631714, + "learning_rate": 0.0003424372431640625, + "loss": 3.751605987548828, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.0066695502027869225, + "learning_rate": 0.0003416198066220997, + "loss": 3.7561278343200684, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.006601935718208551, + "learning_rate": 0.0003408033154513655, + "loss": 3.754878520965576, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.006911450996994972, + "learning_rate": 0.00033998777030969085, + "loss": 3.69533109664917, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.0070345234125852585, + "learning_rate": 0.0003391731718541465, + "loss": 3.7328338623046875, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.007238153368234634, + "learning_rate": 0.00033835952074103964, + "loss": 3.7370123863220215, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.007295276038348675, + "learning_rate": 0.0003375468176259142, + "loss": 3.7123947143554688, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.006471444386988878, + "learning_rate": 0.0003367350631635506, + "loss": 3.73347806930542, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.007136075291782618, + "learning_rate": 0.00033592425800796447, + "loss": 3.7337799072265625, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.0065771411173045635, + "learning_rate": 0.00033511440281240725, + "loss": 3.733506441116333, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.006566311698406935, + "learning_rate": 0.0003343054982293643, + "loss": 3.7325632572174072, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.006389300804585218, + "learning_rate": 0.00033349754491055535, + "loss": 3.7027206420898438, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.006425539031624794, + "learning_rate": 0.00033269054350693386, + "loss": 3.751466989517212, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.006735894829034805, + "learning_rate": 0.0003318844946686865, + "loss": 3.731363296508789, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.0061791036278009415, + "learning_rate": 0.00033107939904523264, + "loss": 3.7592201232910156, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.007923831231892109, + "learning_rate": 0.0003302752572852219, + "loss": 3.7247159481048584, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.371238708496094, + "eval_runtime": 51.312, + "eval_samples_per_second": 47.591, + "eval_steps_per_second": 1.501, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.007161229848861694, + "learning_rate": 0.0003294720700365385, + "loss": 3.6985998153686523, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.007182417903095484, + "learning_rate": 0.00032866983794629457, + "loss": 3.7116494178771973, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.0072847092524170876, + "learning_rate": 0.00032786856166083463, + "loss": 3.6759095191955566, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.006694195792078972, + "learning_rate": 0.0003270682418257323, + "loss": 3.6635122299194336, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.0073003629222512245, + "learning_rate": 0.0003262688790857914, + "loss": 3.7130517959594727, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.006921815685927868, + "learning_rate": 0.0003254704740850432, + "loss": 3.6923139095306396, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.007585546933114529, + "learning_rate": 0.00032467302746674805, + "loss": 3.680659770965576, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.0069264741614460945, + "learning_rate": 0.0003238765398733956, + "loss": 3.7042646408081055, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.007581264711916447, + "learning_rate": 0.0003230810119467003, + "loss": 3.6890318393707275, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.007325596176087856, + "learning_rate": 0.00032228644432760544, + "loss": 3.728785753250122, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.006920192390680313, + "learning_rate": 0.0003214928376562787, + "loss": 3.765744686126709, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.007247452158480883, + "learning_rate": 0.00032070019257211596, + "loss": 3.699038505554199, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.0073222690261900425, + "learning_rate": 0.0003199085097137361, + "loss": 3.6657447814941406, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.0074598658829927444, + "learning_rate": 0.000319117789718984, + "loss": 3.6929068565368652, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.007144949398934841, + "learning_rate": 0.0003183280332249286, + "loss": 3.7166645526885986, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.0068069938570261, + "learning_rate": 0.0003175392408678621, + "loss": 3.650606155395508, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.007320682052522898, + "learning_rate": 0.0003167514132833012, + "loss": 3.69647216796875, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.007024327293038368, + "learning_rate": 0.00031596455110598264, + "loss": 3.6764140129089355, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.007796750869601965, + "learning_rate": 0.0003151786549698689, + "loss": 3.6841444969177246, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.007526555098593235, + "learning_rate": 0.0003143937255081411, + "loss": 3.7181622982025146, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.00689823878929019, + "learning_rate": 0.0003136097633532032, + "loss": 3.6805999279022217, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.007449387572705746, + "learning_rate": 0.00031282676913667903, + "loss": 3.719672679901123, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.0072627803310751915, + "learning_rate": 0.00031204474348941336, + "loss": 3.692244291305542, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.0073951175436377525, + "learning_rate": 0.00031126368704147026, + "loss": 3.734830379486084, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 0.007486097048968077, + "learning_rate": 0.0003104836004221318, + "loss": 3.6959404945373535, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.007235408294945955, + "learning_rate": 0.0003097044842599005, + "loss": 3.714496612548828, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.0074761975556612015, + "learning_rate": 0.00030892633918249584, + "loss": 3.6801066398620605, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.007354153320193291, + "learning_rate": 0.0003081491658168555, + "loss": 3.7053956985473633, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 0.0077071161940693855, + "learning_rate": 0.0003073729647891338, + "loss": 3.724217653274536, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.007286814507097006, + "learning_rate": 0.00030659773672470227, + "loss": 3.717681646347046, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.00741427019238472, + "learning_rate": 0.00030582348224814803, + "loss": 3.7178781032562256, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.007106068544089794, + "learning_rate": 0.00030505020198327355, + "loss": 3.671164035797119, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.006869092117995024, + "learning_rate": 0.0003042778965530969, + "loss": 3.7310664653778076, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.00765156140550971, + "learning_rate": 0.0003035065665798508, + "loss": 3.7417078018188477, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.006932055577635765, + "learning_rate": 0.00030273621268498183, + "loss": 3.6812591552734375, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.008092030882835388, + "learning_rate": 0.00030196683548915055, + "loss": 3.69986891746521, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.00719967158511281, + "learning_rate": 0.00030119843561222915, + "loss": 3.7247209548950195, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.007408039178699255, + "learning_rate": 0.00030043101367330504, + "loss": 3.7106001377105713, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.007386322598904371, + "learning_rate": 0.00029966457029067503, + "loss": 3.7073440551757812, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.007485459093004465, + "learning_rate": 0.00029889910608184924, + "loss": 3.730858325958252, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.007635290268808603, + "learning_rate": 0.00029813462166354814, + "loss": 3.6909053325653076, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0069068302400410175, + "learning_rate": 0.0002973711176517032, + "loss": 3.7286834716796875, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.00698851840570569, + "learning_rate": 0.0002966085946614562, + "loss": 3.6587166786193848, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.007312684785574675, + "learning_rate": 0.0002958470533071569, + "loss": 3.716031789779663, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.006896916776895523, + "learning_rate": 0.00029508649420236687, + "loss": 3.7105724811553955, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.007188099902123213, + "learning_rate": 0.00029432691795985394, + "loss": 3.712202548980713, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.006414467468857765, + "learning_rate": 0.000293568325191595, + "loss": 3.71462345123291, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.006734516471624374, + "learning_rate": 0.00029281071650877496, + "loss": 3.7277159690856934, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.0067634680308401585, + "learning_rate": 0.00029205409252178524, + "loss": 3.6998372077941895, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.006864939816296101, + "learning_rate": 0.00029129845384022446, + "loss": 3.7656338214874268, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.006820365786552429, + "learning_rate": 0.00029054380107289626, + "loss": 3.7018871307373047, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.006954800803214312, + "learning_rate": 0.00028979013482781114, + "loss": 3.7410342693328857, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.0072957612574100494, + "learning_rate": 0.00028903745571218453, + "loss": 3.6964735984802246, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.006805056240409613, + "learning_rate": 0.00028828576433243675, + "loss": 3.6988587379455566, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.007712861057370901, + "learning_rate": 0.0002875350612941908, + "loss": 3.73240327835083, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.006683751940727234, + "learning_rate": 0.0002867853472022755, + "loss": 3.6936707496643066, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.007726171053946018, + "learning_rate": 0.0002860366226607223, + "loss": 3.7266287803649902, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.007187390234321356, + "learning_rate": 0.00028528888827276396, + "loss": 3.7381067276000977, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.007067828439176083, + "learning_rate": 0.00028454214464083706, + "loss": 3.71884822845459, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.007044029422104359, + "learning_rate": 0.00028379639236657966, + "loss": 3.6823341846466064, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.007027678657323122, + "learning_rate": 0.000283051632050831, + "loss": 3.7312417030334473, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.0076786368153989315, + "learning_rate": 0.00028230786429363046, + "loss": 3.775012493133545, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.0067380438558757305, + "learning_rate": 0.000281565089694218, + "loss": 3.7203941345214844, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.007500625215470791, + "learning_rate": 0.00028082330885103516, + "loss": 3.7320876121520996, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.006601814646273851, + "learning_rate": 0.00028008252236172003, + "loss": 3.7289505004882812, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.00755319744348526, + "learning_rate": 0.00027934273082311147, + "loss": 3.7076311111450195, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.0068661835975945, + "learning_rate": 0.0002786039348312459, + "loss": 3.6950671672821045, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.007409265730530024, + "learning_rate": 0.00027786613498135843, + "loss": 3.7798776626586914, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.006714935880154371, + "learning_rate": 0.0002771293318678804, + "loss": 3.690377712249756, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.007498119492083788, + "learning_rate": 0.00027639352608444087, + "loss": 3.741222381591797, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.006700210273265839, + "learning_rate": 0.0002756587182238656, + "loss": 3.7065296173095703, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.008518884889781475, + "learning_rate": 0.00027492490887817597, + "loss": 3.7010321617126465, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 4.372326374053955, + "eval_runtime": 48.97, + "eval_samples_per_second": 49.867, + "eval_steps_per_second": 1.572, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.007463657297194004, + "learning_rate": 0.00027419209863858865, + "loss": 3.6893815994262695, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.0069931019097566605, + "learning_rate": 0.00027346028809551576, + "loss": 3.6620845794677734, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.0076762037351727486, + "learning_rate": 0.00027272947783856433, + "loss": 3.715946674346924, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.007143040653318167, + "learning_rate": 0.0002719996684565341, + "loss": 3.67921781539917, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.007577868644148111, + "learning_rate": 0.00027127086053741964, + "loss": 3.6886801719665527, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.007018061354756355, + "learning_rate": 0.0002705430546684085, + "loss": 3.673877716064453, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.0072300247848033905, + "learning_rate": 0.0002698162514358805, + "loss": 3.730193614959717, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.00714140385389328, + "learning_rate": 0.00026909045142540843, + "loss": 3.6612749099731445, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.0071529485285282135, + "learning_rate": 0.0002683656552217553, + "loss": 3.641969680786133, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.00726350536569953, + "learning_rate": 0.0002676418634088772, + "loss": 3.697068452835083, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.007082062773406506, + "learning_rate": 0.00026691907656991963, + "loss": 3.6732699871063232, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.0069284942001104355, + "learning_rate": 0.00026619729528721897, + "loss": 3.6701955795288086, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.007283426355570555, + "learning_rate": 0.00026547652014230147, + "loss": 3.6422624588012695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.006717352196574211, + "learning_rate": 0.0002647567517158825, + "loss": 3.6600942611694336, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.006842337548732758, + "learning_rate": 0.00026403799058786695, + "loss": 3.690095901489258, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.006624994333833456, + "learning_rate": 0.000263320237337347, + "loss": 3.7145729064941406, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.006822563707828522, + "learning_rate": 0.0002626034925426036, + "loss": 3.734589099884033, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.006343246903270483, + "learning_rate": 0.0002618877567811054, + "loss": 3.676018238067627, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007226140238344669, + "learning_rate": 0.00026117303062950795, + "loss": 3.705630302429199, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.0066663846373558044, + "learning_rate": 0.0002604593146636522, + "loss": 3.707524299621582, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.007017293944954872, + "learning_rate": 0.00025974660945856726, + "loss": 3.654975652694702, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.006494272034615278, + "learning_rate": 0.0002590349155884669, + "loss": 3.7084226608276367, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.007430107332766056, + "learning_rate": 0.0002583242336267492, + "loss": 3.681856393814087, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.0065765841864049435, + "learning_rate": 0.00025761456414599807, + "loss": 3.7148854732513428, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.00703916372731328, + "learning_rate": 0.00025690590771798165, + "loss": 3.707977294921875, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.006881422363221645, + "learning_rate": 0.0002561982649136517, + "loss": 3.6452829837799072, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.006829532328993082, + "learning_rate": 0.00025549163630314233, + "loss": 3.6812632083892822, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.0069242967292666435, + "learning_rate": 0.00025478602245577255, + "loss": 3.6699113845825195, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.00673352275043726, + "learning_rate": 0.00025408142394004265, + "loss": 3.67539644241333, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.006906552705913782, + "learning_rate": 0.0002533778413236342, + "loss": 3.695683479309082, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007096703629940748, + "learning_rate": 0.0002526752751734114, + "loss": 3.720625877380371, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006843909155577421, + "learning_rate": 0.000251973726055419, + "loss": 3.6780426502227783, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.006776071153581142, + "learning_rate": 0.00025127319453488277, + "loss": 3.702991485595703, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.006684580352157354, + "learning_rate": 0.00025057368117620797, + "loss": 3.7161166667938232, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006576182786375284, + "learning_rate": 0.00024987518654297935, + "loss": 3.711947202682495, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.006553746294230223, + "learning_rate": 0.0002491777111979629, + "loss": 3.699580669403076, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.0063622179441154, + "learning_rate": 0.00024848125570310067, + "loss": 3.660463809967041, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006369025446474552, + "learning_rate": 0.00024778582061951466, + "loss": 3.712718963623047, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.0066086845472455025, + "learning_rate": 0.00024709140650750426, + "loss": 3.7337706089019775, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.006698573008179665, + "learning_rate": 0.00024639801392654687, + "loss": 3.693399429321289, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006572197191417217, + "learning_rate": 0.0002457056434352956, + "loss": 3.711644411087036, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.0066736009903252125, + "learning_rate": 0.0002450142955915812, + "loss": 3.706545352935791, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006779580842703581, + "learning_rate": 0.0002443239709524101, + "loss": 3.6896963119506836, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.0064375437796115875, + "learning_rate": 0.0002436346700739646, + "loss": 3.646042823791504, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.006759198382496834, + "learning_rate": 0.00024294639351160198, + "loss": 3.6894006729125977, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.00644648214802146, + "learning_rate": 0.00024225914181985343, + "loss": 3.7088422775268555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.00702979601919651, + "learning_rate": 0.00024157291555242663, + "loss": 3.705204486846924, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006393326912075281, + "learning_rate": 0.00024088771526220045, + "loss": 3.738778829574585, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.006522264331579208, + "learning_rate": 0.00024020354150122884, + "loss": 3.7124788761138916, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.006918660830706358, + "learning_rate": 0.0002395203948207383, + "loss": 3.703273296356201, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.0067405398003757, + "learning_rate": 0.00023883827577112786, + "loss": 3.735112190246582, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.006900794804096222, + "learning_rate": 0.00023815718490196856, + "loss": 3.707573413848877, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.006647802423685789, + "learning_rate": 0.0002374771227620022, + "loss": 3.7194061279296875, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.006786358542740345, + "learning_rate": 0.00023679808989914383, + "loss": 3.684828758239746, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.006440348923206329, + "learning_rate": 0.0002361200868604771, + "loss": 3.679628610610962, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.006847103592008352, + "learning_rate": 0.0002354431141922572, + "loss": 3.7070424556732178, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.00643899803981185, + "learning_rate": 0.00023476717243990815, + "loss": 3.7474443912506104, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.0070899054408073425, + "learning_rate": 0.00023409226214802462, + "loss": 3.6964569091796875, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.006639422383159399, + "learning_rate": 0.00023341838386036994, + "loss": 3.7240355014801025, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.0067732552997767925, + "learning_rate": 0.000232745538119875, + "loss": 3.691311836242676, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.006991941947489977, + "learning_rate": 0.0002320737254686396, + "loss": 3.703491449356079, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.006394840776920319, + "learning_rate": 0.0002314029464479313, + "loss": 3.6963045597076416, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.00672922981902957, + "learning_rate": 0.00023073320159818463, + "loss": 3.6913256645202637, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.006367729511111975, + "learning_rate": 0.00023006449145900022, + "loss": 3.712766170501709, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.006370930001139641, + "learning_rate": 0.0002293968165691467, + "loss": 3.683473587036133, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.007074430584907532, + "learning_rate": 0.00022873017746655758, + "loss": 3.7364859580993652, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.00689726835116744, + "learning_rate": 0.00022806457468833126, + "loss": 3.7360360622406006, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.006516663823276758, + "learning_rate": 0.0002274000087707323, + "loss": 3.7183632850646973, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.006655904930084944, + "learning_rate": 0.00022673648024918938, + "loss": 3.7091434001922607, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.006546393036842346, + "learning_rate": 0.00022607398965829604, + "loss": 3.682645320892334, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.006579853594303131, + "learning_rate": 0.0002254125375318079, + "loss": 3.707092761993408, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.007749224081635475, + "learning_rate": 0.00022475212440264524, + "loss": 3.6835598945617676, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 4.3733673095703125, + "eval_runtime": 54.197, + "eval_samples_per_second": 45.058, + "eval_steps_per_second": 1.421, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.006405573338270187, + "learning_rate": 0.000224092750802892, + "loss": 3.7344212532043457, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.006732994224876165, + "learning_rate": 0.0002234344172637924, + "loss": 3.684095621109009, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.00679958239197731, + "learning_rate": 0.00022277712431575434, + "loss": 3.697983741760254, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.00670752814039588, + "learning_rate": 0.00022212087248834695, + "loss": 3.6364173889160156, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.006356659810990095, + "learning_rate": 0.00022146566231030071, + "loss": 3.642988681793213, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.006624842528253794, + "learning_rate": 0.000220811494309506, + "loss": 3.6682677268981934, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.006486820988357067, + "learning_rate": 0.00022015836901301408, + "loss": 3.6703057289123535, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.006477998569607735, + "learning_rate": 0.0002195062869470374, + "loss": 3.6891942024230957, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.006716342642903328, + "learning_rate": 0.0002188552486369454, + "loss": 3.6837644577026367, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.006492446642369032, + "learning_rate": 0.00021820525460726888, + "loss": 3.66949462890625, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.0064978403970599174, + "learning_rate": 0.00021755630538169502, + "loss": 3.656477451324463, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.006488895509392023, + "learning_rate": 0.000216908401483072, + "loss": 3.683609962463379, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.0063864924013614655, + "learning_rate": 0.00021626154343340316, + "loss": 3.6635069847106934, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.006423276849091053, + "learning_rate": 0.00021561573175385077, + "loss": 3.675956964492798, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.006375681143254042, + "learning_rate": 0.00021497096696473364, + "loss": 3.662991523742676, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.006157599855214357, + "learning_rate": 0.00021432724958552707, + "loss": 3.685154438018799, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.0064014773815870285, + "learning_rate": 0.0002136845801348629, + "loss": 3.6482927799224854, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.006111831869930029, + "learning_rate": 0.0002130429591305271, + "loss": 3.6458590030670166, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.0062011126428842545, + "learning_rate": 0.0002124023870894639, + "loss": 3.6651253700256348, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.006603785324841738, + "learning_rate": 0.00021176286452776909, + "loss": 3.6931943893432617, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.006363562773913145, + "learning_rate": 0.00021112439196069516, + "loss": 3.7063207626342773, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.006258734501898289, + "learning_rate": 0.0002104869699026478, + "loss": 3.689711093902588, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.006544198840856552, + "learning_rate": 0.00020985059886718659, + "loss": 3.6803956031799316, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.006236413959413767, + "learning_rate": 0.00020921527936702418, + "loss": 3.681662082672119, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.006620924919843674, + "learning_rate": 0.00020858101191402557, + "loss": 3.6708180904388428, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.006312021054327488, + "learning_rate": 0.00020794779701920895, + "loss": 3.68668270111084, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.006595146376639605, + "learning_rate": 0.00020731563519274407, + "loss": 3.709224224090576, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.006226524710655212, + "learning_rate": 0.00020668452694395234, + "loss": 3.688267707824707, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.006482335273176432, + "learning_rate": 0.00020605447278130616, + "loss": 3.6906940937042236, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.006389188580214977, + "learning_rate": 0.00020542547321242892, + "loss": 3.6813313961029053, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.006554862484335899, + "learning_rate": 0.00020479752874409457, + "loss": 3.645631790161133, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.00646898802369833, + "learning_rate": 0.0002041706398822257, + "loss": 3.691580295562744, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.006872115656733513, + "learning_rate": 0.0002035448071318961, + "loss": 3.716787815093994, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.006251885090023279, + "learning_rate": 0.00020292003099732744, + "loss": 3.666078567504883, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.006771737243980169, + "learning_rate": 0.00020229631198189084, + "loss": 3.621438503265381, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.00606488948687911, + "learning_rate": 0.0002016736505881058, + "loss": 3.6710901260375977, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.006203590892255306, + "learning_rate": 0.00020105204731763832, + "loss": 3.6667685508728027, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.006625751499086618, + "learning_rate": 0.00020043150267130412, + "loss": 3.648806095123291, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.006358640734106302, + "learning_rate": 0.000199812017149064, + "loss": 3.7056827545166016, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.006539507769048214, + "learning_rate": 0.00019919359125002647, + "loss": 3.6906323432922363, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.007050348911434412, + "learning_rate": 0.00019857622547244627, + "loss": 3.6898245811462402, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.006209646817296743, + "learning_rate": 0.00019795992031372358, + "loss": 3.6794214248657227, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.0064559150487184525, + "learning_rate": 0.00019734467627040487, + "loss": 3.6937336921691895, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.006774613168090582, + "learning_rate": 0.00019673049383818025, + "loss": 3.7005720138549805, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.0061691561713814735, + "learning_rate": 0.00019611737351188642, + "loss": 3.7170605659484863, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.0070815603248775005, + "learning_rate": 0.0001955053157855027, + "loss": 3.722456693649292, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.006411026697605848, + "learning_rate": 0.00019489432115215308, + "loss": 3.669548511505127, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.006817718036472797, + "learning_rate": 0.00019428439010410492, + "loss": 3.705488681793213, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.006276566535234451, + "learning_rate": 0.0001936755231327688, + "loss": 3.7110562324523926, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.006713696289807558, + "learning_rate": 0.0001930677207286977, + "loss": 3.7030649185180664, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.006331400014460087, + "learning_rate": 0.00019246098338158676, + "loss": 3.6906652450561523, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.007149284239858389, + "learning_rate": 0.00019185531158027384, + "loss": 3.7055468559265137, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.006395523902028799, + "learning_rate": 0.00019125070581273734, + "loss": 3.72355318069458, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.007234807591885328, + "learning_rate": 0.0001906471665660978, + "loss": 3.6685009002685547, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.006079623010009527, + "learning_rate": 0.00019004469432661478, + "loss": 3.7051684856414795, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.00699756620451808, + "learning_rate": 0.0001894432895796907, + "loss": 3.6713948249816895, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.006799036171287298, + "learning_rate": 0.00018884295280986531, + "loss": 3.7098617553710938, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.006591275800019503, + "learning_rate": 0.0001882436845008197, + "loss": 3.685445785522461, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.006709378212690353, + "learning_rate": 0.0001876454851353733, + "loss": 3.6773548126220703, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.00651182746514678, + "learning_rate": 0.0001870483551954847, + "loss": 3.6529507637023926, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.006494257599115372, + "learning_rate": 0.00018645229516225093, + "loss": 3.692331552505493, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.006487370003014803, + "learning_rate": 0.00018585730551590619, + "loss": 3.679516315460205, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.006413135677576065, + "learning_rate": 0.00018526338673582307, + "loss": 3.707998752593994, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.006509249564260244, + "learning_rate": 0.00018467053930051156, + "loss": 3.7103686332702637, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.006331913638859987, + "learning_rate": 0.0001840787636876177, + "loss": 3.6807055473327637, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.006528888829052448, + "learning_rate": 0.00018348806037392473, + "loss": 3.7196922302246094, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.006212258245795965, + "learning_rate": 0.0001828984298353515, + "loss": 3.716118335723877, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.006641864310950041, + "learning_rate": 0.000182309872546953, + "loss": 3.6726081371307373, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.006433268543332815, + "learning_rate": 0.0001817223889829186, + "loss": 3.690727710723877, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.00644147302955389, + "learning_rate": 0.00018113597961657364, + "loss": 3.6998791694641113, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.006477390881627798, + "learning_rate": 0.00018055064492037722, + "loss": 3.700925350189209, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.0077974810265004635, + "learning_rate": 0.00017996638536592322, + "loss": 3.735599994659424, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 4.376750946044922, + "eval_runtime": 52.3576, + "eval_samples_per_second": 46.641, + "eval_steps_per_second": 1.471, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.006609050091356039, + "learning_rate": 0.00017938320142393897, + "loss": 3.6405727863311768, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.006470015738159418, + "learning_rate": 0.00017880109356428522, + "loss": 3.6639411449432373, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.006597487255930901, + "learning_rate": 0.00017822006225595596, + "loss": 3.6797537803649902, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.0064637972973287106, + "learning_rate": 0.0001776401079670772, + "loss": 3.6798789501190186, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.006872601341456175, + "learning_rate": 0.00017706123116490775, + "loss": 3.649874210357666, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.0064028468914330006, + "learning_rate": 0.00017648343231583825, + "loss": 3.671036720275879, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.006735685281455517, + "learning_rate": 0.00017590671188539084, + "loss": 3.6818368434906006, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.006652928423136473, + "learning_rate": 0.00017533107033821895, + "loss": 3.660362720489502, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.006564146373420954, + "learning_rate": 0.00017475650813810598, + "loss": 3.6444082260131836, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.006476754322648048, + "learning_rate": 0.00017418302574796716, + "loss": 3.6836254596710205, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.007108788937330246, + "learning_rate": 0.0001736106236298462, + "loss": 3.6424341201782227, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.006570066325366497, + "learning_rate": 0.00017303930224491735, + "loss": 3.663379192352295, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.0066614942625164986, + "learning_rate": 0.00017246906205348395, + "loss": 3.6918721199035645, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.00704583153128624, + "learning_rate": 0.00017189990351497824, + "loss": 3.692559242248535, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.006617981940507889, + "learning_rate": 0.00017133182708796126, + "loss": 3.6591439247131348, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.006916946265846491, + "learning_rate": 0.00017076483323012127, + "loss": 3.7165699005126953, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.006832561455667019, + "learning_rate": 0.00017019892239827534, + "loss": 3.7023603916168213, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.00654558464884758, + "learning_rate": 0.00016963409504836777, + "loss": 3.6702237129211426, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.007000794634222984, + "learning_rate": 0.0001690703516354698, + "loss": 3.69088077545166, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.006473604589700699, + "learning_rate": 0.00016850769261377856, + "loss": 3.6641716957092285, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.006787394639104605, + "learning_rate": 0.000167946118436619, + "loss": 3.6577839851379395, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.006220053881406784, + "learning_rate": 0.00016738562955644144, + "loss": 3.6735315322875977, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.006513087544590235, + "learning_rate": 0.00016682622642482113, + "loss": 3.624807119369507, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.006552594713866711, + "learning_rate": 0.0001662679094924592, + "loss": 3.679218053817749, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.006950687617063522, + "learning_rate": 0.0001657106792091814, + "loss": 3.6864023208618164, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.006092227529734373, + "learning_rate": 0.0001651545360239387, + "loss": 3.693777084350586, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.006924188230186701, + "learning_rate": 0.0001645994803848046, + "loss": 3.68534779548645, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.0067315176129341125, + "learning_rate": 0.00016404551273897826, + "loss": 3.632850170135498, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.007006707601249218, + "learning_rate": 0.00016349263353278146, + "loss": 3.6664111614227295, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.0067031108774244785, + "learning_rate": 0.00016294084321165846, + "loss": 3.6863315105438232, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.006502983160316944, + "learning_rate": 0.00016239014222017682, + "loss": 3.678041458129883, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.006409413646906614, + "learning_rate": 0.00016184053100202665, + "loss": 3.692004919052124, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.006251887418329716, + "learning_rate": 0.0001612920100000198, + "loss": 3.64532470703125, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.0065657407976686954, + "learning_rate": 0.00016074457965608942, + "loss": 3.638624429702759, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.006246988195925951, + "learning_rate": 0.00016019824041129004, + "loss": 3.668254852294922, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.0065050385892391205, + "learning_rate": 0.0001596529927057984, + "loss": 3.6901469230651855, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.006240491755306721, + "learning_rate": 0.00015910883697890977, + "loss": 3.6796071529388428, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.006674987263977528, + "learning_rate": 0.00015856577366904075, + "loss": 3.7001953125, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.006260441616177559, + "learning_rate": 0.00015802380321372776, + "loss": 3.6646556854248047, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.006696469150483608, + "learning_rate": 0.00015748292604962687, + "loss": 3.6791043281555176, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.006716041825711727, + "learning_rate": 0.00015694314261251245, + "loss": 3.710442066192627, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.006856406573206186, + "learning_rate": 0.0001564044533372785, + "loss": 3.6615681648254395, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.0064330073073506355, + "learning_rate": 0.00015586685865793748, + "loss": 3.707085609436035, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.006748909596353769, + "learning_rate": 0.00015533035900761955, + "loss": 3.6814656257629395, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.006306119728833437, + "learning_rate": 0.00015479495481857282, + "loss": 3.6595420837402344, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.0061928280629217625, + "learning_rate": 0.00015426064652216255, + "loss": 3.686880111694336, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.006499494891613722, + "learning_rate": 0.00015372743454887173, + "loss": 3.6901583671569824, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.005914116278290749, + "learning_rate": 0.00015319531932829917, + "loss": 3.699162006378174, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.006499269045889378, + "learning_rate": 0.00015266430128916075, + "loss": 3.6740684509277344, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.006433961447328329, + "learning_rate": 0.000152134380859288, + "loss": 3.684006929397583, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.006057603284716606, + "learning_rate": 0.00015160555846562826, + "loss": 3.6612191200256348, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.006337927188724279, + "learning_rate": 0.00015107783453424456, + "loss": 3.6981983184814453, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.006348243448883295, + "learning_rate": 0.00015055120949031365, + "loss": 3.6667895317077637, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.0060755726881325245, + "learning_rate": 0.00015002568375812876, + "loss": 3.6844797134399414, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.006059548351913691, + "learning_rate": 0.0001495012577610959, + "loss": 3.6676461696624756, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.006432189140468836, + "learning_rate": 0.00014897793192173602, + "loss": 3.7036523818969727, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.006043825298547745, + "learning_rate": 0.00014845570666168249, + "loss": 3.7043371200561523, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.0066383229568600655, + "learning_rate": 0.0001479345824016837, + "loss": 3.6830334663391113, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.006384526379406452, + "learning_rate": 0.00014741455956160005, + "loss": 3.6746675968170166, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.006477308925241232, + "learning_rate": 0.0001468956385604038, + "loss": 3.6788225173950195, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.006130026653409004, + "learning_rate": 0.00014637781981618054, + "loss": 3.6821796894073486, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.006257957778871059, + "learning_rate": 0.00014586110374612777, + "loss": 3.683966636657715, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.006265794392675161, + "learning_rate": 0.00014534549076655425, + "loss": 3.691389560699463, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.006294097751379013, + "learning_rate": 0.0001448309812928794, + "loss": 3.659594774246216, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.006432943977415562, + "learning_rate": 0.00014431757573963504, + "loss": 3.665633201599121, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.006102072540670633, + "learning_rate": 0.0001438052745204626, + "loss": 3.6348228454589844, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.00636800192296505, + "learning_rate": 0.00014329407804811333, + "loss": 3.624040365219116, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.006060190033167601, + "learning_rate": 0.00014278398673444966, + "loss": 3.6986045837402344, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.0061101666651666164, + "learning_rate": 0.00014227500099044254, + "loss": 3.6816675662994385, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.006048897281289101, + "learning_rate": 0.00014176712122617293, + "loss": 3.6739861965179443, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.005954945459961891, + "learning_rate": 0.00014126034785082997, + "loss": 3.679354190826416, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.007545145694166422, + "learning_rate": 0.0001407546812727119, + "loss": 3.7448782920837402, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 4.377684116363525, + "eval_runtime": 51.0591, + "eval_samples_per_second": 47.827, + "eval_steps_per_second": 1.508, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.00643246341496706, + "learning_rate": 0.00014025012189922562, + "loss": 3.6544783115386963, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.006309461314231157, + "learning_rate": 0.00013974667013688495, + "loss": 3.701963186264038, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.006177262403070927, + "learning_rate": 0.00013924432639131215, + "loss": 3.649641275405884, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.006319771986454725, + "learning_rate": 0.00013874309106723636, + "loss": 3.67417573928833, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.006548433564603329, + "learning_rate": 0.00013824296456849426, + "loss": 3.6257686614990234, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.006148583721369505, + "learning_rate": 0.00013774394729802804, + "loss": 3.675339460372925, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.0063438634388148785, + "learning_rate": 0.00013724603965788716, + "loss": 3.666316270828247, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.006417148746550083, + "learning_rate": 0.00013674924204922732, + "loss": 3.6692237854003906, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.0062498473562300205, + "learning_rate": 0.00013625355487230872, + "loss": 3.70363187789917, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.006249656435102224, + "learning_rate": 0.00013575897852649796, + "loss": 3.691108226776123, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.006240040063858032, + "learning_rate": 0.0001352655134102657, + "loss": 3.6781342029571533, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.006186297629028559, + "learning_rate": 0.00013477315992118885, + "loss": 3.644334316253662, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.006283515132963657, + "learning_rate": 0.00013428191845594675, + "loss": 3.6691460609436035, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.0064705004915595055, + "learning_rate": 0.00013379178941032427, + "loss": 3.64168119430542, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.006210929248481989, + "learning_rate": 0.00013330277317920959, + "loss": 3.66428804397583, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.006095333956182003, + "learning_rate": 0.00013281487015659395, + "loss": 3.6205215454101562, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.006364993751049042, + "learning_rate": 0.00013232808073557244, + "loss": 3.662804126739502, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.006166066508740187, + "learning_rate": 0.00013184240530834173, + "loss": 3.702338218688965, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.006147229578346014, + "learning_rate": 0.00013135784426620282, + "loss": 3.660149097442627, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.006086066365242004, + "learning_rate": 0.00013087439799955674, + "loss": 3.652472972869873, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.006371276918798685, + "learning_rate": 0.00013039206689790815, + "loss": 3.6471242904663086, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.005952878389507532, + "learning_rate": 0.000129910851349862, + "loss": 3.6450729370117188, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.006385558284819126, + "learning_rate": 0.00012943075174312536, + "loss": 3.6730473041534424, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.005896324757486582, + "learning_rate": 0.00012895176846450582, + "loss": 3.6646409034729004, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.006053723394870758, + "learning_rate": 0.00012847390189991121, + "loss": 3.6568682193756104, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.005929332692176104, + "learning_rate": 0.00012799715243435027, + "loss": 3.7030997276306152, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.006153058726340532, + "learning_rate": 0.00012752152045193116, + "loss": 3.6949849128723145, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.006193236447870731, + "learning_rate": 0.0001270470063358624, + "loss": 3.6040501594543457, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.005849961191415787, + "learning_rate": 0.0001265736104684509, + "loss": 3.602717399597168, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.0060405549593269825, + "learning_rate": 0.00012610133323110356, + "loss": 3.668717861175537, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.0062002381309866905, + "learning_rate": 0.00012563017500432565, + "loss": 3.6333389282226562, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.006093314848840237, + "learning_rate": 0.00012516013616772065, + "loss": 3.662696361541748, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.005765238776803017, + "learning_rate": 0.00012469121709999026, + "loss": 3.6843791007995605, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.005989165045320988, + "learning_rate": 0.00012422341817893448, + "loss": 3.6904735565185547, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.006150389555841684, + "learning_rate": 0.00012375673978145022, + "loss": 3.6614508628845215, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.005898299627006054, + "learning_rate": 0.00012329118228353178, + "loss": 3.7193222045898438, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.0058402069844305515, + "learning_rate": 0.00012282674606027016, + "loss": 3.669564723968506, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.006015103310346603, + "learning_rate": 0.00012236343148585413, + "loss": 3.686734914779663, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.005990280769765377, + "learning_rate": 0.00012190123893356703, + "loss": 3.664945125579834, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.006110248621553183, + "learning_rate": 0.00012144016877578942, + "loss": 3.6546223163604736, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.006006988696753979, + "learning_rate": 0.000120980221383997, + "loss": 3.669175386428833, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.006008700001984835, + "learning_rate": 0.00012052139712876141, + "loss": 3.672668695449829, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.005862198304384947, + "learning_rate": 0.00012006369637974875, + "loss": 3.6824991703033447, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.005960221868008375, + "learning_rate": 0.00011960711950572005, + "loss": 3.6258697509765625, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.005813480354845524, + "learning_rate": 0.00011915166687453174, + "loss": 3.6463029384613037, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.006002446636557579, + "learning_rate": 0.0001186973388531332, + "loss": 3.721970558166504, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.006029803771525621, + "learning_rate": 0.00011824413580756844, + "loss": 3.685335159301758, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.005933200940489769, + "learning_rate": 0.0001177920581029752, + "loss": 3.6834492683410645, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.005942870397120714, + "learning_rate": 0.00011734110610358423, + "loss": 3.670149326324463, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.005972386337816715, + "learning_rate": 0.00011689128017271901, + "loss": 3.6798110008239746, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.005964178126305342, + "learning_rate": 0.00011644258067279662, + "loss": 3.6620044708251953, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.0059594810009002686, + "learning_rate": 0.00011599500796532605, + "loss": 3.683058738708496, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.005901646800339222, + "learning_rate": 0.00011554856241090844, + "loss": 3.6791443824768066, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.006178564857691526, + "learning_rate": 0.00011510324436923708, + "loss": 3.6620421409606934, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.005973632447421551, + "learning_rate": 0.0001146590541990962, + "loss": 3.6821961402893066, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.005914160516113043, + "learning_rate": 0.00011421599225836265, + "loss": 3.6588668823242188, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.005887249484658241, + "learning_rate": 0.00011377405890400266, + "loss": 3.683187484741211, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.006068149581551552, + "learning_rate": 0.00011333325449207415, + "loss": 3.6633477210998535, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.005748802796006203, + "learning_rate": 0.00011289357937772547, + "loss": 3.676607608795166, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.006032158620655537, + "learning_rate": 0.00011245503391519489, + "loss": 3.652369499206543, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.00615714630112052, + "learning_rate": 0.00011201761845781049, + "loss": 3.7210354804992676, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.0059432960115373135, + "learning_rate": 0.00011158133335799003, + "loss": 3.640336036682129, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.0058724163100123405, + "learning_rate": 0.00011114617896724026, + "loss": 3.638444185256958, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.006073266267776489, + "learning_rate": 0.00011071215563615752, + "loss": 3.680121421813965, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.005904203746467829, + "learning_rate": 0.00011027926371442662, + "loss": 3.6883950233459473, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.005889971274882555, + "learning_rate": 0.00010984750355082045, + "loss": 3.6875810623168945, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.005878197029232979, + "learning_rate": 0.00010941687549320068, + "loss": 3.6341497898101807, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.005862570833414793, + "learning_rate": 0.00010898737988851677, + "loss": 3.648092746734619, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.0058378661051392555, + "learning_rate": 0.00010855901708280482, + "loss": 3.627786636352539, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0057954080402851105, + "learning_rate": 0.0001081317874211895, + "loss": 3.636265754699707, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.005763230379670858, + "learning_rate": 0.00010770569124788201, + "loss": 3.6991324424743652, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.007435750216245651, + "learning_rate": 0.00010728072890618029, + "loss": 3.698052406311035, + "step": 3240 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.3813260125995008e+18, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-3240/training_args.bin b/runs/baseline/checkpoint-3240/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-3240/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-360/chat_template.jinja b/runs/baseline/checkpoint-360/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-360/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-360/config.json b/runs/baseline/checkpoint-360/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-360/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-360/generation_config.json b/runs/baseline/checkpoint-360/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-360/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-360/model.safetensors b/runs/baseline/checkpoint-360/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5db2a911183563f3ac12920bdb93a1e3b04d2886 --- /dev/null +++ b/runs/baseline/checkpoint-360/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e278725cdeece35f0bd1ccc41b8ceef8ec080ca3c6adb12629b41bab70293699 +size 583356232 diff --git a/runs/baseline/checkpoint-360/optimizer.pt b/runs/baseline/checkpoint-360/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..5225a05ef9671f4864b140f4c2629b84e4791604 --- /dev/null +++ b/runs/baseline/checkpoint-360/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f656fc0021dd1568063069ead3d1c685d1d893d5bd6dc43f598982bd9201f93 +size 1166825803 diff --git a/runs/baseline/checkpoint-360/rng_state_0.pth b/runs/baseline/checkpoint-360/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..3a2bb760d0b0b50f4b7b6980626fda172a127f68 --- /dev/null +++ b/runs/baseline/checkpoint-360/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9627a7b92eeed978894313d8642baf0ecdcd808f2102543d8b58136102ad5997 +size 14917 diff --git a/runs/baseline/checkpoint-360/rng_state_1.pth b/runs/baseline/checkpoint-360/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..7b307a1ab4859d8d838e7d20293e5eb15985549a --- /dev/null +++ b/runs/baseline/checkpoint-360/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d79c621c2f337ae714092486c08d5a224dca6ddc4285fb0014e9b648d0c9cae +size 14917 diff --git a/runs/baseline/checkpoint-360/scheduler.pt b/runs/baseline/checkpoint-360/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..3acc96860dc2d45c77cec0d2f55bac5383c2a389 --- /dev/null +++ b/runs/baseline/checkpoint-360/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c2cb27444ef7fe125f0634d6f0887d628ac22fdc5ea5d28ad2b2519c38009cf +size 1465 diff --git a/runs/baseline/checkpoint-360/tokenizer.json b/runs/baseline/checkpoint-360/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-360/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-360/tokenizer_config.json b/runs/baseline/checkpoint-360/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-360/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-360/trainer_state.json b/runs/baseline/checkpoint-360/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..a8bbe99d48af8155557780ea09d06ee9992b0a43 --- /dev/null +++ b/runs/baseline/checkpoint-360/trainer_state.json @@ -0,0 +1,2586 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 360, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.534806680666112e+17, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-360/training_args.bin b/runs/baseline/checkpoint-360/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-360/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-3600/chat_template.jinja b/runs/baseline/checkpoint-3600/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-3600/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-3600/config.json b/runs/baseline/checkpoint-3600/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-3600/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-3600/generation_config.json b/runs/baseline/checkpoint-3600/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-3600/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-3600/model.safetensors b/runs/baseline/checkpoint-3600/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3f5e82e28e3d58b9343dd1945b5a105f60539327 --- /dev/null +++ b/runs/baseline/checkpoint-3600/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72ce76048e8d4b7c3387c7d5b90f6274666606ec81c5f33ae4591fc0e12177dc +size 583356232 diff --git a/runs/baseline/checkpoint-3600/optimizer.pt b/runs/baseline/checkpoint-3600/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..f229dcb5a60c628efe4c5e718334ddd1f2a29801 --- /dev/null +++ b/runs/baseline/checkpoint-3600/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fe36266830f9e17f5f5a3ffc66f118b6e2bed37f92d7ff3bf9956544a04fc42 +size 1166825803 diff --git a/runs/baseline/checkpoint-3600/rng_state_0.pth b/runs/baseline/checkpoint-3600/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..42e5fb760a4884c24f3899e0e2c34f9ab7d35d5e --- /dev/null +++ b/runs/baseline/checkpoint-3600/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9abac844de24b211d8b88d44e1866f089edee417850520a358afe34f866a821 +size 14917 diff --git a/runs/baseline/checkpoint-3600/rng_state_1.pth b/runs/baseline/checkpoint-3600/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..d6f10c7be65a25152763463abf33eb700d784288 --- /dev/null +++ b/runs/baseline/checkpoint-3600/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68ba205d24a50bf8cdf4686014dfa7b5557eb04aefa63431dced1c3d75cbacab +size 14917 diff --git a/runs/baseline/checkpoint-3600/scheduler.pt b/runs/baseline/checkpoint-3600/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..f967f3272dea4dcd87d1b7218164c5b8c405a0a2 --- /dev/null +++ b/runs/baseline/checkpoint-3600/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96be326da54197356c9c3f871e031619329af2df05a743623e3ef304de77920c +size 1465 diff --git a/runs/baseline/checkpoint-3600/tokenizer.json b/runs/baseline/checkpoint-3600/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-3600/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-3600/tokenizer_config.json b/runs/baseline/checkpoint-3600/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-3600/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-3600/trainer_state.json b/runs/baseline/checkpoint-3600/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..c80990d5a85c7660d3d9ce92b20c3d5aa2c6ea4a --- /dev/null +++ b/runs/baseline/checkpoint-3600/trainer_state.json @@ -0,0 +1,25626 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 50.0, + "eval_steps": 500, + "global_step": 3600, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + }, + { + "epoch": 10.0, + "eval_loss": 5.602636814117432, + "eval_runtime": 54.6199, + "eval_samples_per_second": 44.709, + "eval_steps_per_second": 1.41, + "step": 720 + }, + { + "epoch": 10.013973799126637, + "grad_norm": 0.02089761756360531, + "learning_rate": 0.0027759185760599576, + "loss": 5.461747646331787, + "step": 721 + }, + { + "epoch": 10.027947598253276, + "grad_norm": 0.018181778490543365, + "learning_rate": 0.0027752139775442278, + "loss": 5.492741584777832, + "step": 722 + }, + { + "epoch": 10.041921397379912, + "grad_norm": 0.020166238769888878, + "learning_rate": 0.0027745083636968575, + "loss": 5.476003646850586, + "step": 723 + }, + { + "epoch": 10.055895196506551, + "grad_norm": 0.015491694211959839, + "learning_rate": 0.0027738017350863488, + "loss": 5.425764560699463, + "step": 724 + }, + { + "epoch": 10.069868995633188, + "grad_norm": 0.022291697561740875, + "learning_rate": 0.0027730940922820182, + "loss": 5.45866584777832, + "step": 725 + }, + { + "epoch": 10.083842794759825, + "grad_norm": 0.018074683845043182, + "learning_rate": 0.002772385435854002, + "loss": 5.465143203735352, + "step": 726 + }, + { + "epoch": 10.097816593886463, + "grad_norm": 0.016872268170118332, + "learning_rate": 0.002771675766373251, + "loss": 5.446003437042236, + "step": 727 + }, + { + "epoch": 10.1117903930131, + "grad_norm": 0.023103339597582817, + "learning_rate": 0.0027709650844115336, + "loss": 5.411036491394043, + "step": 728 + }, + { + "epoch": 10.125764192139737, + "grad_norm": 0.027957158163189888, + "learning_rate": 0.0027702533905414324, + "loss": 5.452442169189453, + "step": 729 + }, + { + "epoch": 10.139737991266376, + "grad_norm": 0.019935399293899536, + "learning_rate": 0.002769540685336348, + "loss": 5.412145614624023, + "step": 730 + }, + { + "epoch": 10.153711790393013, + "grad_norm": 0.014351371675729752, + "learning_rate": 0.0027688269693704926, + "loss": 5.434524059295654, + "step": 731 + }, + { + "epoch": 10.167685589519651, + "grad_norm": 0.0152254244312644, + "learning_rate": 0.0027681122432188948, + "loss": 5.428544998168945, + "step": 732 + }, + { + "epoch": 10.181659388646288, + "grad_norm": 0.015071330592036247, + "learning_rate": 0.0027673965074573965, + "loss": 5.459929466247559, + "step": 733 + }, + { + "epoch": 10.195633187772925, + "grad_norm": 0.023415958508849144, + "learning_rate": 0.0027666797626626537, + "loss": 5.447159767150879, + "step": 734 + }, + { + "epoch": 10.209606986899564, + "grad_norm": 0.017228802666068077, + "learning_rate": 0.0027659620094121337, + "loss": 5.442273139953613, + "step": 735 + }, + { + "epoch": 10.2235807860262, + "grad_norm": 0.014184504747390747, + "learning_rate": 0.002765243248284118, + "loss": 5.448305130004883, + "step": 736 + }, + { + "epoch": 10.237554585152838, + "grad_norm": 0.024384526535868645, + "learning_rate": 0.0027645234798576988, + "loss": 5.4076337814331055, + "step": 737 + }, + { + "epoch": 10.251528384279476, + "grad_norm": 0.043152231723070145, + "learning_rate": 0.0027638027047127813, + "loss": 5.45344352722168, + "step": 738 + }, + { + "epoch": 10.265502183406113, + "grad_norm": 0.02889758162200451, + "learning_rate": 0.002763080923430081, + "loss": 5.489194393157959, + "step": 739 + }, + { + "epoch": 10.279475982532752, + "grad_norm": 0.02686968818306923, + "learning_rate": 0.002762358136591123, + "loss": 5.493659496307373, + "step": 740 + }, + { + "epoch": 10.293449781659389, + "grad_norm": 0.020467868074774742, + "learning_rate": 0.002761634344778245, + "loss": 5.410489559173584, + "step": 741 + }, + { + "epoch": 10.307423580786025, + "grad_norm": 0.019759224727749825, + "learning_rate": 0.002760909548574592, + "loss": 5.439385890960693, + "step": 742 + }, + { + "epoch": 10.321397379912664, + "grad_norm": 0.0233843382447958, + "learning_rate": 0.0027601837485641194, + "loss": 5.445036888122559, + "step": 743 + }, + { + "epoch": 10.335371179039301, + "grad_norm": 0.031753458082675934, + "learning_rate": 0.0027594569453315916, + "loss": 5.503895282745361, + "step": 744 + }, + { + "epoch": 10.34934497816594, + "grad_norm": 0.02089349366724491, + "learning_rate": 0.002758729139462581, + "loss": 5.45689582824707, + "step": 745 + }, + { + "epoch": 10.363318777292577, + "grad_norm": 0.03339357674121857, + "learning_rate": 0.002758000331543466, + "loss": 5.522361755371094, + "step": 746 + }, + { + "epoch": 10.377292576419213, + "grad_norm": 0.019328996539115906, + "learning_rate": 0.002757270522161436, + "loss": 5.441373825073242, + "step": 747 + }, + { + "epoch": 10.391266375545852, + "grad_norm": 0.019863856956362724, + "learning_rate": 0.002756539711904484, + "loss": 5.4633283615112305, + "step": 748 + }, + { + "epoch": 10.405240174672489, + "grad_norm": 0.02141754701733589, + "learning_rate": 0.0027558079013614116, + "loss": 5.453215599060059, + "step": 749 + }, + { + "epoch": 10.419213973799126, + "grad_norm": 0.016591915860772133, + "learning_rate": 0.002755075091121824, + "loss": 5.463696479797363, + "step": 750 + }, + { + "epoch": 10.433187772925764, + "grad_norm": 0.01955028623342514, + "learning_rate": 0.0027543412817761345, + "loss": 5.411355018615723, + "step": 751 + }, + { + "epoch": 10.447161572052401, + "grad_norm": 0.016791662201285362, + "learning_rate": 0.0027536064739155596, + "loss": 5.459614276885986, + "step": 752 + }, + { + "epoch": 10.46113537117904, + "grad_norm": 0.019661981612443924, + "learning_rate": 0.00275287066813212, + "loss": 5.440699577331543, + "step": 753 + }, + { + "epoch": 10.475109170305677, + "grad_norm": 0.01927965320646763, + "learning_rate": 0.002752133865018642, + "loss": 5.409865856170654, + "step": 754 + }, + { + "epoch": 10.489082969432314, + "grad_norm": 0.016483580693602562, + "learning_rate": 0.002751396065168754, + "loss": 5.390999794006348, + "step": 755 + }, + { + "epoch": 10.503056768558952, + "grad_norm": 0.016264773905277252, + "learning_rate": 0.0027506572691768888, + "loss": 5.418543815612793, + "step": 756 + }, + { + "epoch": 10.51703056768559, + "grad_norm": 0.01958460919559002, + "learning_rate": 0.00274991747763828, + "loss": 5.403289794921875, + "step": 757 + }, + { + "epoch": 10.531004366812226, + "grad_norm": 0.01861695572733879, + "learning_rate": 0.002749176691148965, + "loss": 5.431911468505859, + "step": 758 + }, + { + "epoch": 10.544978165938865, + "grad_norm": 0.029594263061881065, + "learning_rate": 0.002748434910305782, + "loss": 5.399242401123047, + "step": 759 + }, + { + "epoch": 10.558951965065502, + "grad_norm": 0.028574613854289055, + "learning_rate": 0.0027476921357063697, + "loss": 5.440446853637695, + "step": 760 + }, + { + "epoch": 10.57292576419214, + "grad_norm": 0.028612585738301277, + "learning_rate": 0.0027469483679491693, + "loss": 5.410346031188965, + "step": 761 + }, + { + "epoch": 10.586899563318777, + "grad_norm": 0.020704898983240128, + "learning_rate": 0.00274620360763342, + "loss": 5.44694709777832, + "step": 762 + }, + { + "epoch": 10.600873362445414, + "grad_norm": 0.020542364567518234, + "learning_rate": 0.002745457855359163, + "loss": 5.402653694152832, + "step": 763 + }, + { + "epoch": 10.614847161572053, + "grad_norm": 0.015762662515044212, + "learning_rate": 0.0027447111117272363, + "loss": 5.396594047546387, + "step": 764 + }, + { + "epoch": 10.62882096069869, + "grad_norm": 0.022096984088420868, + "learning_rate": 0.002743963377339278, + "loss": 5.398744106292725, + "step": 765 + }, + { + "epoch": 10.642794759825328, + "grad_norm": 0.018791764974594116, + "learning_rate": 0.002743214652797724, + "loss": 5.397758483886719, + "step": 766 + }, + { + "epoch": 10.656768558951965, + "grad_norm": 0.01642526686191559, + "learning_rate": 0.0027424649387058093, + "loss": 5.411459922790527, + "step": 767 + }, + { + "epoch": 10.670742358078602, + "grad_norm": 0.017931703478097916, + "learning_rate": 0.0027417142356675637, + "loss": 5.390264511108398, + "step": 768 + }, + { + "epoch": 10.68471615720524, + "grad_norm": 0.018819602206349373, + "learning_rate": 0.0027409625442878153, + "loss": 5.407539367675781, + "step": 769 + }, + { + "epoch": 10.698689956331878, + "grad_norm": 0.019389411434531212, + "learning_rate": 0.0027402098651721886, + "loss": 5.386474609375, + "step": 770 + }, + { + "epoch": 10.712663755458514, + "grad_norm": 0.013645983301103115, + "learning_rate": 0.002739456198927104, + "loss": 5.373845100402832, + "step": 771 + }, + { + "epoch": 10.726637554585153, + "grad_norm": 0.013846895657479763, + "learning_rate": 0.002738701546159776, + "loss": 5.387155055999756, + "step": 772 + }, + { + "epoch": 10.74061135371179, + "grad_norm": 0.013553125783801079, + "learning_rate": 0.002737945907478215, + "loss": 5.377801895141602, + "step": 773 + }, + { + "epoch": 10.754585152838429, + "grad_norm": 0.012771612964570522, + "learning_rate": 0.002737189283491225, + "loss": 5.344066143035889, + "step": 774 + }, + { + "epoch": 10.768558951965066, + "grad_norm": 0.020465301349759102, + "learning_rate": 0.002736431674808405, + "loss": 5.382771968841553, + "step": 775 + }, + { + "epoch": 10.782532751091702, + "grad_norm": 0.014798106625676155, + "learning_rate": 0.0027356730820401465, + "loss": 5.370522499084473, + "step": 776 + }, + { + "epoch": 10.796506550218341, + "grad_norm": 0.01566634140908718, + "learning_rate": 0.0027349135057976332, + "loss": 5.351980209350586, + "step": 777 + }, + { + "epoch": 10.810480349344978, + "grad_norm": 0.017485102638602257, + "learning_rate": 0.002734152946692843, + "loss": 5.355443477630615, + "step": 778 + }, + { + "epoch": 10.824454148471617, + "grad_norm": 0.016646096482872963, + "learning_rate": 0.0027333914053385444, + "loss": 5.34829044342041, + "step": 779 + }, + { + "epoch": 10.838427947598253, + "grad_norm": 0.017293352633714676, + "learning_rate": 0.0027326288823482964, + "loss": 5.3519487380981445, + "step": 780 + }, + { + "epoch": 10.85240174672489, + "grad_norm": 0.018873002380132675, + "learning_rate": 0.002731865378336452, + "loss": 5.348442077636719, + "step": 781 + }, + { + "epoch": 10.866375545851529, + "grad_norm": 0.015504784882068634, + "learning_rate": 0.002731100893918151, + "loss": 5.341490268707275, + "step": 782 + }, + { + "epoch": 10.880349344978166, + "grad_norm": 0.014270870015025139, + "learning_rate": 0.0027303354297093254, + "loss": 5.359251022338867, + "step": 783 + }, + { + "epoch": 10.894323144104803, + "grad_norm": 0.018820680677890778, + "learning_rate": 0.002729568986326695, + "loss": 5.390280723571777, + "step": 784 + }, + { + "epoch": 10.908296943231441, + "grad_norm": 0.021008187904953957, + "learning_rate": 0.0027288015643877708, + "loss": 5.338380336761475, + "step": 785 + }, + { + "epoch": 10.922270742358078, + "grad_norm": 0.016266655176877975, + "learning_rate": 0.0027280331645108498, + "loss": 5.347043037414551, + "step": 786 + }, + { + "epoch": 10.936244541484717, + "grad_norm": 0.01831045374274254, + "learning_rate": 0.0027272637873150185, + "loss": 5.326537132263184, + "step": 787 + }, + { + "epoch": 10.950218340611354, + "grad_norm": 0.02073794975876808, + "learning_rate": 0.002726493433420149, + "loss": 5.3607025146484375, + "step": 788 + }, + { + "epoch": 10.96419213973799, + "grad_norm": 0.022534754127264023, + "learning_rate": 0.0027257221034469033, + "loss": 5.353292465209961, + "step": 789 + }, + { + "epoch": 10.97816593886463, + "grad_norm": 0.021627362817525864, + "learning_rate": 0.002724949798016727, + "loss": 5.373323917388916, + "step": 790 + }, + { + "epoch": 10.992139737991266, + "grad_norm": 0.018226243555545807, + "learning_rate": 0.0027241765177518523, + "loss": 5.344578266143799, + "step": 791 + }, + { + "epoch": 11.0, + "grad_norm": 0.02032875455915928, + "learning_rate": 0.002723402263275298, + "loss": 5.329159259796143, + "step": 792 + }, + { + "epoch": 11.0, + "eval_loss": 5.44969367980957, + "eval_runtime": 51.2294, + "eval_samples_per_second": 47.668, + "eval_steps_per_second": 1.503, + "step": 792 + }, + { + "epoch": 11.013973799126637, + "grad_norm": 0.01613701693713665, + "learning_rate": 0.0027226270352108664, + "loss": 5.3024678230285645, + "step": 793 + }, + { + "epoch": 11.027947598253276, + "grad_norm": 0.01940973661839962, + "learning_rate": 0.0027218508341831445, + "loss": 5.270181179046631, + "step": 794 + }, + { + "epoch": 11.041921397379912, + "grad_norm": 0.015840493142604828, + "learning_rate": 0.002721073660817504, + "loss": 5.261706352233887, + "step": 795 + }, + { + "epoch": 11.055895196506551, + "grad_norm": 0.017390675842761993, + "learning_rate": 0.0027202955157401, + "loss": 5.262207984924316, + "step": 796 + }, + { + "epoch": 11.069868995633188, + "grad_norm": 0.015184012241661549, + "learning_rate": 0.0027195163995778686, + "loss": 5.297155380249023, + "step": 797 + }, + { + "epoch": 11.083842794759825, + "grad_norm": 0.013417073525488377, + "learning_rate": 0.00271873631295853, + "loss": 5.279838562011719, + "step": 798 + }, + { + "epoch": 11.097816593886463, + "grad_norm": 0.013024317100644112, + "learning_rate": 0.0027179552565105864, + "loss": 5.265446186065674, + "step": 799 + }, + { + "epoch": 11.1117903930131, + "grad_norm": 0.01426338218152523, + "learning_rate": 0.002717173230863321, + "loss": 5.284180164337158, + "step": 800 + }, + { + "epoch": 11.125764192139737, + "grad_norm": 0.019604040309786797, + "learning_rate": 0.002716390236646797, + "loss": 5.286281108856201, + "step": 801 + }, + { + "epoch": 11.139737991266376, + "grad_norm": 0.019039476290345192, + "learning_rate": 0.0027156062744918587, + "loss": 5.290433406829834, + "step": 802 + }, + { + "epoch": 11.153711790393013, + "grad_norm": 0.019284334033727646, + "learning_rate": 0.0027148213450301313, + "loss": 5.262433052062988, + "step": 803 + }, + { + "epoch": 11.167685589519651, + "grad_norm": 0.0145154669880867, + "learning_rate": 0.0027140354488940175, + "loss": 5.24387264251709, + "step": 804 + }, + { + "epoch": 11.181659388646288, + "grad_norm": 0.014776417054235935, + "learning_rate": 0.0027132485867166992, + "loss": 5.282435417175293, + "step": 805 + }, + { + "epoch": 11.195633187772925, + "grad_norm": 0.012711729854345322, + "learning_rate": 0.0027124607591321377, + "loss": 5.246735572814941, + "step": 806 + }, + { + "epoch": 11.209606986899564, + "grad_norm": 0.013394835405051708, + "learning_rate": 0.0027116719667750716, + "loss": 5.270739555358887, + "step": 807 + }, + { + "epoch": 11.2235807860262, + "grad_norm": 0.013468850404024124, + "learning_rate": 0.002710882210281016, + "loss": 5.27161169052124, + "step": 808 + }, + { + "epoch": 11.237554585152838, + "grad_norm": 0.030262840911746025, + "learning_rate": 0.002710091490286264, + "loss": 5.256673812866211, + "step": 809 + }, + { + "epoch": 11.251528384279476, + "grad_norm": 0.013560288585722446, + "learning_rate": 0.0027092998074278846, + "loss": 5.253674507141113, + "step": 810 + }, + { + "epoch": 11.265502183406113, + "grad_norm": 0.014592813327908516, + "learning_rate": 0.002708507162343721, + "loss": 5.25804328918457, + "step": 811 + }, + { + "epoch": 11.279475982532752, + "grad_norm": 0.014328169636428356, + "learning_rate": 0.0027077135556723953, + "loss": 5.235117435455322, + "step": 812 + }, + { + "epoch": 11.293449781659389, + "grad_norm": 0.01277366653084755, + "learning_rate": 0.0027069189880533, + "loss": 5.2053985595703125, + "step": 813 + }, + { + "epoch": 11.307423580786025, + "grad_norm": 0.012948785908520222, + "learning_rate": 0.0027061234601266043, + "loss": 5.214653015136719, + "step": 814 + }, + { + "epoch": 11.321397379912664, + "grad_norm": 0.014892828650772572, + "learning_rate": 0.002705326972533252, + "loss": 5.227547645568848, + "step": 815 + }, + { + "epoch": 11.335371179039301, + "grad_norm": 0.01427230704575777, + "learning_rate": 0.002704529525914957, + "loss": 5.241065502166748, + "step": 816 + }, + { + "epoch": 11.34934497816594, + "grad_norm": 0.01650116965174675, + "learning_rate": 0.0027037311209142092, + "loss": 5.227713108062744, + "step": 817 + }, + { + "epoch": 11.363318777292577, + "grad_norm": 0.019728578627109528, + "learning_rate": 0.0027029317581742677, + "loss": 5.2624921798706055, + "step": 818 + }, + { + "epoch": 11.377292576419213, + "grad_norm": 0.023847149685025215, + "learning_rate": 0.0027021314383391657, + "loss": 5.278811454772949, + "step": 819 + }, + { + "epoch": 11.391266375545852, + "grad_norm": 0.020690469071269035, + "learning_rate": 0.0027013301620537056, + "loss": 5.238919258117676, + "step": 820 + }, + { + "epoch": 11.405240174672489, + "grad_norm": 0.021211959421634674, + "learning_rate": 0.0027005279299634616, + "loss": 5.253684997558594, + "step": 821 + }, + { + "epoch": 11.419213973799126, + "grad_norm": 0.025854922831058502, + "learning_rate": 0.002699724742714778, + "loss": 5.2538652420043945, + "step": 822 + }, + { + "epoch": 11.433187772925764, + "grad_norm": 0.0215485580265522, + "learning_rate": 0.0026989206009547678, + "loss": 5.224540710449219, + "step": 823 + }, + { + "epoch": 11.447161572052401, + "grad_norm": 0.019827792420983315, + "learning_rate": 0.0026981155053313137, + "loss": 5.220674514770508, + "step": 824 + }, + { + "epoch": 11.46113537117904, + "grad_norm": 0.017689093947410583, + "learning_rate": 0.002697309456493066, + "loss": 5.228511810302734, + "step": 825 + }, + { + "epoch": 11.475109170305677, + "grad_norm": 0.022564806044101715, + "learning_rate": 0.002696502455089445, + "loss": 5.295933723449707, + "step": 826 + }, + { + "epoch": 11.489082969432314, + "grad_norm": 0.020070280879735947, + "learning_rate": 0.0026956945017706357, + "loss": 5.230920791625977, + "step": 827 + }, + { + "epoch": 11.503056768558952, + "grad_norm": 0.025744661688804626, + "learning_rate": 0.002694885597187593, + "loss": 5.296453952789307, + "step": 828 + }, + { + "epoch": 11.51703056768559, + "grad_norm": 0.021263007074594498, + "learning_rate": 0.0026940757419920357, + "loss": 5.220481872558594, + "step": 829 + }, + { + "epoch": 11.531004366812226, + "grad_norm": 0.017577245831489563, + "learning_rate": 0.0026932649368364495, + "loss": 5.231488227844238, + "step": 830 + }, + { + "epoch": 11.544978165938865, + "grad_norm": 0.015562832355499268, + "learning_rate": 0.002692453182374086, + "loss": 5.175224304199219, + "step": 831 + }, + { + "epoch": 11.558951965065502, + "grad_norm": 0.014115403406322002, + "learning_rate": 0.0026916404792589606, + "loss": 5.228387832641602, + "step": 832 + }, + { + "epoch": 11.57292576419214, + "grad_norm": 0.012863830663263798, + "learning_rate": 0.002690826828145854, + "loss": 5.186727523803711, + "step": 833 + }, + { + "epoch": 11.586899563318777, + "grad_norm": 0.014151126146316528, + "learning_rate": 0.002690012229690309, + "loss": 5.1694817543029785, + "step": 834 + }, + { + "epoch": 11.600873362445414, + "grad_norm": 0.01316932775080204, + "learning_rate": 0.0026891966845486353, + "loss": 5.19931697845459, + "step": 835 + }, + { + "epoch": 11.614847161572053, + "grad_norm": 0.011340726166963577, + "learning_rate": 0.0026883801933779003, + "loss": 5.238619804382324, + "step": 836 + }, + { + "epoch": 11.62882096069869, + "grad_norm": 0.012616089545190334, + "learning_rate": 0.0026875627568359378, + "loss": 5.183376789093018, + "step": 837 + }, + { + "epoch": 11.642794759825328, + "grad_norm": 0.012274038046598434, + "learning_rate": 0.002686744375581341, + "loss": 5.199946403503418, + "step": 838 + }, + { + "epoch": 11.656768558951965, + "grad_norm": 0.014990398660302162, + "learning_rate": 0.002685925050273465, + "loss": 5.168451309204102, + "step": 839 + }, + { + "epoch": 11.670742358078602, + "grad_norm": 0.011399757117033005, + "learning_rate": 0.0026851047815724255, + "loss": 5.173828125, + "step": 840 + }, + { + "epoch": 11.68471615720524, + "grad_norm": 0.013638390228152275, + "learning_rate": 0.0026842835701390982, + "loss": 5.142702102661133, + "step": 841 + }, + { + "epoch": 11.698689956331878, + "grad_norm": 0.014472718350589275, + "learning_rate": 0.0026834614166351177, + "loss": 5.173436164855957, + "step": 842 + }, + { + "epoch": 11.712663755458514, + "grad_norm": 0.014509808272123337, + "learning_rate": 0.002682638321722879, + "loss": 5.173975467681885, + "step": 843 + }, + { + "epoch": 11.726637554585153, + "grad_norm": 0.013576583005487919, + "learning_rate": 0.0026818142860655352, + "loss": 5.1435089111328125, + "step": 844 + }, + { + "epoch": 11.74061135371179, + "grad_norm": 0.012914472259581089, + "learning_rate": 0.0026809893103269954, + "loss": 5.151381492614746, + "step": 845 + }, + { + "epoch": 11.754585152838429, + "grad_norm": 0.013803352601826191, + "learning_rate": 0.0026801633951719288, + "loss": 5.133159637451172, + "step": 846 + }, + { + "epoch": 11.768558951965066, + "grad_norm": 0.018488213419914246, + "learning_rate": 0.0026793365412657606, + "loss": 5.183903694152832, + "step": 847 + }, + { + "epoch": 11.782532751091702, + "grad_norm": 0.01843969337642193, + "learning_rate": 0.002678508749274671, + "loss": 5.1660356521606445, + "step": 848 + }, + { + "epoch": 11.796506550218341, + "grad_norm": 0.017181839793920517, + "learning_rate": 0.0026776800198655983, + "loss": 5.179847717285156, + "step": 849 + }, + { + "epoch": 11.810480349344978, + "grad_norm": 0.0148160420358181, + "learning_rate": 0.0026768503537062333, + "loss": 5.149198532104492, + "step": 850 + }, + { + "epoch": 11.824454148471617, + "grad_norm": 0.013251296244561672, + "learning_rate": 0.0026760197514650247, + "loss": 5.192000389099121, + "step": 851 + }, + { + "epoch": 11.838427947598253, + "grad_norm": 0.010746927931904793, + "learning_rate": 0.0026751882138111717, + "loss": 5.138680458068848, + "step": 852 + }, + { + "epoch": 11.85240174672489, + "grad_norm": 0.012316994369029999, + "learning_rate": 0.002674355741414631, + "loss": 5.169965744018555, + "step": 853 + }, + { + "epoch": 11.866375545851529, + "grad_norm": 0.01191797386854887, + "learning_rate": 0.0026735223349461106, + "loss": 5.128267288208008, + "step": 854 + }, + { + "epoch": 11.880349344978166, + "grad_norm": 0.013525084592401981, + "learning_rate": 0.0026726879950770698, + "loss": 5.124795436859131, + "step": 855 + }, + { + "epoch": 11.894323144104803, + "grad_norm": 0.014843370765447617, + "learning_rate": 0.002671852722479721, + "loss": 5.1497087478637695, + "step": 856 + }, + { + "epoch": 11.908296943231441, + "grad_norm": 0.015285412780940533, + "learning_rate": 0.00267101651782703, + "loss": 5.093118190765381, + "step": 857 + }, + { + "epoch": 11.922270742358078, + "grad_norm": 0.013555832207202911, + "learning_rate": 0.0026701793817927107, + "loss": 5.13249397277832, + "step": 858 + }, + { + "epoch": 11.936244541484717, + "grad_norm": 0.010569742880761623, + "learning_rate": 0.0026693413150512273, + "loss": 5.084077835083008, + "step": 859 + }, + { + "epoch": 11.950218340611354, + "grad_norm": 0.011483813636004925, + "learning_rate": 0.0026685023182777975, + "loss": 5.128049850463867, + "step": 860 + }, + { + "epoch": 11.96419213973799, + "grad_norm": 0.013724366202950478, + "learning_rate": 0.0026676623921483844, + "loss": 5.157539367675781, + "step": 861 + }, + { + "epoch": 11.97816593886463, + "grad_norm": 0.014337432570755482, + "learning_rate": 0.0026668215373397016, + "loss": 5.149948596954346, + "step": 862 + }, + { + "epoch": 11.992139737991266, + "grad_norm": 0.014101454056799412, + "learning_rate": 0.00266597975452921, + "loss": 5.117250442504883, + "step": 863 + }, + { + "epoch": 12.0, + "grad_norm": 0.016004368662834167, + "learning_rate": 0.0026651370443951197, + "loss": 5.143221855163574, + "step": 864 + }, + { + "epoch": 12.0, + "eval_loss": 5.249833583831787, + "eval_runtime": 51.5543, + "eval_samples_per_second": 47.368, + "eval_steps_per_second": 1.494, + "step": 864 + }, + { + "epoch": 12.013973799126637, + "grad_norm": 0.01297403872013092, + "learning_rate": 0.0026642934076163867, + "loss": 5.089287281036377, + "step": 865 + }, + { + "epoch": 12.027947598253276, + "grad_norm": 0.012867718935012817, + "learning_rate": 0.0026634488448727144, + "loss": 5.0784101486206055, + "step": 866 + }, + { + "epoch": 12.041921397379912, + "grad_norm": 0.014600927010178566, + "learning_rate": 0.0026626033568445504, + "loss": 5.050220489501953, + "step": 867 + }, + { + "epoch": 12.055895196506551, + "grad_norm": 0.016619572415947914, + "learning_rate": 0.0026617569442130905, + "loss": 5.079303741455078, + "step": 868 + }, + { + "epoch": 12.069868995633188, + "grad_norm": 0.015560587868094444, + "learning_rate": 0.0026609096076602736, + "loss": 5.072385311126709, + "step": 869 + }, + { + "epoch": 12.083842794759825, + "grad_norm": 0.01774057187139988, + "learning_rate": 0.0026600613478687833, + "loss": 5.122529983520508, + "step": 870 + }, + { + "epoch": 12.097816593886463, + "grad_norm": 0.019971400499343872, + "learning_rate": 0.0026592121655220472, + "loss": 5.085966110229492, + "step": 871 + }, + { + "epoch": 12.1117903930131, + "grad_norm": 0.017667256295681, + "learning_rate": 0.0026583620613042365, + "loss": 5.078432559967041, + "step": 872 + }, + { + "epoch": 12.125764192139737, + "grad_norm": 0.022369660437107086, + "learning_rate": 0.002657511035900264, + "loss": 5.0630903244018555, + "step": 873 + }, + { + "epoch": 12.139737991266376, + "grad_norm": 0.017964458093047142, + "learning_rate": 0.002656659089995787, + "loss": 5.0846076011657715, + "step": 874 + }, + { + "epoch": 12.153711790393013, + "grad_norm": 0.021769145503640175, + "learning_rate": 0.002655806224277201, + "loss": 5.081344127655029, + "step": 875 + }, + { + "epoch": 12.167685589519651, + "grad_norm": 0.02487620897591114, + "learning_rate": 0.0026549524394316464, + "loss": 5.106287956237793, + "step": 876 + }, + { + "epoch": 12.181659388646288, + "grad_norm": 0.022516218945384026, + "learning_rate": 0.0026540977361470012, + "loss": 5.128098487854004, + "step": 877 + }, + { + "epoch": 12.195633187772925, + "grad_norm": 0.02099926583468914, + "learning_rate": 0.0026532421151118845, + "loss": 5.103562355041504, + "step": 878 + }, + { + "epoch": 12.209606986899564, + "grad_norm": 0.01899387687444687, + "learning_rate": 0.002652385577015655, + "loss": 5.062675476074219, + "step": 879 + }, + { + "epoch": 12.2235807860262, + "grad_norm": 0.016801996156573296, + "learning_rate": 0.002651528122548409, + "loss": 5.096002578735352, + "step": 880 + }, + { + "epoch": 12.237554585152838, + "grad_norm": 0.014009200967848301, + "learning_rate": 0.0026506697524009836, + "loss": 5.075201034545898, + "step": 881 + }, + { + "epoch": 12.251528384279476, + "grad_norm": 0.014653858728706837, + "learning_rate": 0.002649810467264951, + "loss": 5.090478897094727, + "step": 882 + }, + { + "epoch": 12.265502183406113, + "grad_norm": 0.014918150380253792, + "learning_rate": 0.0026489502678326226, + "loss": 5.070797920227051, + "step": 883 + }, + { + "epoch": 12.279475982532752, + "grad_norm": 0.01327288057655096, + "learning_rate": 0.0026480891547970443, + "loss": 5.030500411987305, + "step": 884 + }, + { + "epoch": 12.293449781659389, + "grad_norm": 0.012885140255093575, + "learning_rate": 0.002647227128852, + "loss": 5.0845537185668945, + "step": 885 + }, + { + "epoch": 12.307423580786025, + "grad_norm": 0.012848512269556522, + "learning_rate": 0.002646364190692009, + "loss": 5.079484939575195, + "step": 886 + }, + { + "epoch": 12.321397379912664, + "grad_norm": 0.01149789709597826, + "learning_rate": 0.002645500341012324, + "loss": 5.036316394805908, + "step": 887 + }, + { + "epoch": 12.335371179039301, + "grad_norm": 0.010360757820308208, + "learning_rate": 0.0026446355805089337, + "loss": 5.04129695892334, + "step": 888 + }, + { + "epoch": 12.34934497816594, + "grad_norm": 0.010426290333271027, + "learning_rate": 0.00264376990987856, + "loss": 5.051270484924316, + "step": 889 + }, + { + "epoch": 12.363318777292577, + "grad_norm": 0.01018153689801693, + "learning_rate": 0.002642903329818658, + "loss": 5.102342128753662, + "step": 890 + }, + { + "epoch": 12.377292576419213, + "grad_norm": 0.011879283003509045, + "learning_rate": 0.0026420358410274153, + "loss": 5.053501129150391, + "step": 891 + }, + { + "epoch": 12.391266375545852, + "grad_norm": 0.010418329387903214, + "learning_rate": 0.002641167444203752, + "loss": 5.030653476715088, + "step": 892 + }, + { + "epoch": 12.405240174672489, + "grad_norm": 0.012302013114094734, + "learning_rate": 0.0026402981400473202, + "loss": 5.0402069091796875, + "step": 893 + }, + { + "epoch": 12.419213973799126, + "grad_norm": 0.011392916552722454, + "learning_rate": 0.002639427929258502, + "loss": 5.024111747741699, + "step": 894 + }, + { + "epoch": 12.433187772925764, + "grad_norm": 0.013075058348476887, + "learning_rate": 0.0026385568125384114, + "loss": 5.021731376647949, + "step": 895 + }, + { + "epoch": 12.447161572052401, + "grad_norm": 0.015349832363426685, + "learning_rate": 0.0026376847905888903, + "loss": 5.03192138671875, + "step": 896 + }, + { + "epoch": 12.46113537117904, + "grad_norm": 0.01588309183716774, + "learning_rate": 0.0026368118641125113, + "loss": 4.966363906860352, + "step": 897 + }, + { + "epoch": 12.475109170305677, + "grad_norm": 0.014378590509295464, + "learning_rate": 0.0026359380338125764, + "loss": 5.070773124694824, + "step": 898 + }, + { + "epoch": 12.489082969432314, + "grad_norm": 0.012306655757129192, + "learning_rate": 0.002635063300393114, + "loss": 5.025704860687256, + "step": 899 + }, + { + "epoch": 12.503056768558952, + "grad_norm": 0.01212004292756319, + "learning_rate": 0.002634187664558881, + "loss": 4.965104103088379, + "step": 900 + }, + { + "epoch": 12.51703056768559, + "grad_norm": 0.013538849540054798, + "learning_rate": 0.0026333111270153624, + "loss": 5.029810905456543, + "step": 901 + }, + { + "epoch": 12.531004366812226, + "grad_norm": 0.012052906677126884, + "learning_rate": 0.0026324336884687672, + "loss": 5.0262908935546875, + "step": 902 + }, + { + "epoch": 12.544978165938865, + "grad_norm": 0.011368660256266594, + "learning_rate": 0.002631555349626033, + "loss": 5.001485824584961, + "step": 903 + }, + { + "epoch": 12.558951965065502, + "grad_norm": 0.010215395130217075, + "learning_rate": 0.0026306761111948205, + "loss": 5.019739151000977, + "step": 904 + }, + { + "epoch": 12.57292576419214, + "grad_norm": 0.011449878104031086, + "learning_rate": 0.0026297959738835167, + "loss": 4.9816083908081055, + "step": 905 + }, + { + "epoch": 12.586899563318777, + "grad_norm": 0.012381811626255512, + "learning_rate": 0.0026289149384012326, + "loss": 5.025032043457031, + "step": 906 + }, + { + "epoch": 12.600873362445414, + "grad_norm": 0.00974641740322113, + "learning_rate": 0.002628033005457802, + "loss": 5.000177383422852, + "step": 907 + }, + { + "epoch": 12.614847161572053, + "grad_norm": 0.009515874087810516, + "learning_rate": 0.0026271501757637833, + "loss": 4.996979713439941, + "step": 908 + }, + { + "epoch": 12.62882096069869, + "grad_norm": 0.011125455610454082, + "learning_rate": 0.0026262664500304555, + "loss": 4.996796607971191, + "step": 909 + }, + { + "epoch": 12.642794759825328, + "grad_norm": 0.013290468603372574, + "learning_rate": 0.0026253818289698207, + "loss": 5.053703308105469, + "step": 910 + }, + { + "epoch": 12.656768558951965, + "grad_norm": 0.014650283381342888, + "learning_rate": 0.002624496313294602, + "loss": 5.001321792602539, + "step": 911 + }, + { + "epoch": 12.670742358078602, + "grad_norm": 0.014537827111780643, + "learning_rate": 0.002623609903718244, + "loss": 5.024584770202637, + "step": 912 + }, + { + "epoch": 12.68471615720524, + "grad_norm": 0.012908566743135452, + "learning_rate": 0.0026227226009549105, + "loss": 5.042878150939941, + "step": 913 + }, + { + "epoch": 12.698689956331878, + "grad_norm": 0.01446808222681284, + "learning_rate": 0.002621834405719485, + "loss": 5.020857810974121, + "step": 914 + }, + { + "epoch": 12.712663755458514, + "grad_norm": 0.01535153016448021, + "learning_rate": 0.0026209453187275708, + "loss": 5.0056304931640625, + "step": 915 + }, + { + "epoch": 12.726637554585153, + "grad_norm": 0.01815616339445114, + "learning_rate": 0.0026200553406954892, + "loss": 5.0467376708984375, + "step": 916 + }, + { + "epoch": 12.74061135371179, + "grad_norm": 0.023135516792535782, + "learning_rate": 0.0026191644723402795, + "loss": 5.033995628356934, + "step": 917 + }, + { + "epoch": 12.754585152838429, + "grad_norm": 0.019682930782437325, + "learning_rate": 0.002618272714379698, + "loss": 5.008149147033691, + "step": 918 + }, + { + "epoch": 12.768558951965066, + "grad_norm": 0.018521904945373535, + "learning_rate": 0.002617380067532217, + "loss": 5.007076263427734, + "step": 919 + }, + { + "epoch": 12.782532751091702, + "grad_norm": 0.02118663489818573, + "learning_rate": 0.0026164865325170278, + "loss": 5.05519962310791, + "step": 920 + }, + { + "epoch": 12.796506550218341, + "grad_norm": 0.024684788659214973, + "learning_rate": 0.0026155921100540344, + "loss": 5.027571201324463, + "step": 921 + }, + { + "epoch": 12.810480349344978, + "grad_norm": 0.02609819918870926, + "learning_rate": 0.0026146968008638567, + "loss": 5.10361385345459, + "step": 922 + }, + { + "epoch": 12.824454148471617, + "grad_norm": 0.025958053767681122, + "learning_rate": 0.00261380060566783, + "loss": 5.0911455154418945, + "step": 923 + }, + { + "epoch": 12.838427947598253, + "grad_norm": 0.017067518085241318, + "learning_rate": 0.002612903525188002, + "loss": 5.045783042907715, + "step": 924 + }, + { + "epoch": 12.85240174672489, + "grad_norm": 0.016510486602783203, + "learning_rate": 0.0026120055601471345, + "loss": 5.069464683532715, + "step": 925 + }, + { + "epoch": 12.866375545851529, + "grad_norm": 0.015552333556115627, + "learning_rate": 0.0026111067112687012, + "loss": 5.024248123168945, + "step": 926 + }, + { + "epoch": 12.880349344978166, + "grad_norm": 0.013432685285806656, + "learning_rate": 0.0026102069792768886, + "loss": 5.004837989807129, + "step": 927 + }, + { + "epoch": 12.894323144104803, + "grad_norm": 0.012300201691687107, + "learning_rate": 0.0026093063648965943, + "loss": 5.031093597412109, + "step": 928 + }, + { + "epoch": 12.908296943231441, + "grad_norm": 0.01197744719684124, + "learning_rate": 0.002608404868853428, + "loss": 4.965237617492676, + "step": 929 + }, + { + "epoch": 12.922270742358078, + "grad_norm": 0.01100246049463749, + "learning_rate": 0.002607502491873708, + "loss": 5.04719352722168, + "step": 930 + }, + { + "epoch": 12.936244541484717, + "grad_norm": 0.010111447423696518, + "learning_rate": 0.0026065992346844626, + "loss": 4.986410617828369, + "step": 931 + }, + { + "epoch": 12.950218340611354, + "grad_norm": 0.010669790208339691, + "learning_rate": 0.002605695098013431, + "loss": 4.9825663566589355, + "step": 932 + }, + { + "epoch": 12.96419213973799, + "grad_norm": 0.010805477388203144, + "learning_rate": 0.0026047900825890593, + "loss": 4.991506576538086, + "step": 933 + }, + { + "epoch": 12.97816593886463, + "grad_norm": 0.010886074975132942, + "learning_rate": 0.002603884189140502, + "loss": 4.987531661987305, + "step": 934 + }, + { + "epoch": 12.992139737991266, + "grad_norm": 0.010735839605331421, + "learning_rate": 0.0026029774183976206, + "loss": 4.982756614685059, + "step": 935 + }, + { + "epoch": 13.0, + "grad_norm": 0.011092593893408775, + "learning_rate": 0.002602069771090985, + "loss": 4.994824409484863, + "step": 936 + }, + { + "epoch": 13.0, + "eval_loss": 5.13364315032959, + "eval_runtime": 50.4195, + "eval_samples_per_second": 48.434, + "eval_steps_per_second": 1.527, + "step": 936 + }, + { + "epoch": 13.013973799126637, + "grad_norm": 0.01005905494093895, + "learning_rate": 0.0026011612479518694, + "loss": 4.918412685394287, + "step": 937 + }, + { + "epoch": 13.027947598253276, + "grad_norm": 0.009904682636260986, + "learning_rate": 0.002600251849712255, + "loss": 4.924630165100098, + "step": 938 + }, + { + "epoch": 13.041921397379912, + "grad_norm": 0.009543103165924549, + "learning_rate": 0.0025993415771048273, + "loss": 4.940309524536133, + "step": 939 + }, + { + "epoch": 13.055895196506551, + "grad_norm": 0.011031470261514187, + "learning_rate": 0.0025984304308629765, + "loss": 4.924609184265137, + "step": 940 + }, + { + "epoch": 13.069868995633188, + "grad_norm": 0.009511022828519344, + "learning_rate": 0.002597518411720796, + "loss": 4.9102325439453125, + "step": 941 + }, + { + "epoch": 13.083842794759825, + "grad_norm": 0.00886621605604887, + "learning_rate": 0.0025966055204130847, + "loss": 4.9236907958984375, + "step": 942 + }, + { + "epoch": 13.097816593886463, + "grad_norm": 0.010057039558887482, + "learning_rate": 0.002595691757675341, + "loss": 4.908633232116699, + "step": 943 + }, + { + "epoch": 13.1117903930131, + "grad_norm": 0.01098555512726307, + "learning_rate": 0.0025947771242437684, + "loss": 4.935927391052246, + "step": 944 + }, + { + "epoch": 13.125764192139737, + "grad_norm": 0.008800056762993336, + "learning_rate": 0.0025938616208552694, + "loss": 4.873291969299316, + "step": 945 + }, + { + "epoch": 13.139737991266376, + "grad_norm": 0.009747766889631748, + "learning_rate": 0.0025929452482474497, + "loss": 4.954436302185059, + "step": 946 + }, + { + "epoch": 13.153711790393013, + "grad_norm": 0.010785684920847416, + "learning_rate": 0.002592028007158613, + "loss": 4.924927711486816, + "step": 947 + }, + { + "epoch": 13.167685589519651, + "grad_norm": 0.012620717287063599, + "learning_rate": 0.0025911098983277645, + "loss": 4.953606605529785, + "step": 948 + }, + { + "epoch": 13.181659388646288, + "grad_norm": 0.014428738504648209, + "learning_rate": 0.002590190922494607, + "loss": 4.951706886291504, + "step": 949 + }, + { + "epoch": 13.195633187772925, + "grad_norm": 0.014496048912405968, + "learning_rate": 0.0025892710803995444, + "loss": 4.941249847412109, + "step": 950 + }, + { + "epoch": 13.209606986899564, + "grad_norm": 0.0109238987788558, + "learning_rate": 0.0025883503727836764, + "loss": 4.939975261688232, + "step": 951 + }, + { + "epoch": 13.2235807860262, + "grad_norm": 0.012271072715520859, + "learning_rate": 0.0025874288003887996, + "loss": 4.880942344665527, + "step": 952 + }, + { + "epoch": 13.237554585152838, + "grad_norm": 0.016302887350320816, + "learning_rate": 0.0025865063639574086, + "loss": 4.931602478027344, + "step": 953 + }, + { + "epoch": 13.251528384279476, + "grad_norm": 0.016137251630425453, + "learning_rate": 0.0025855830642326945, + "loss": 4.918013095855713, + "step": 954 + }, + { + "epoch": 13.265502183406113, + "grad_norm": 0.014732961542904377, + "learning_rate": 0.002584658901958542, + "loss": 4.927577495574951, + "step": 955 + }, + { + "epoch": 13.279475982532752, + "grad_norm": 0.016387946903705597, + "learning_rate": 0.0025837338778795325, + "loss": 4.930819034576416, + "step": 956 + }, + { + "epoch": 13.293449781659389, + "grad_norm": 0.017146151512861252, + "learning_rate": 0.002582807992740941, + "loss": 4.928377151489258, + "step": 957 + }, + { + "epoch": 13.307423580786025, + "grad_norm": 0.016153613105416298, + "learning_rate": 0.0025818812472887375, + "loss": 4.975607395172119, + "step": 958 + }, + { + "epoch": 13.321397379912664, + "grad_norm": 0.013386047445237637, + "learning_rate": 0.0025809536422695825, + "loss": 4.9434709548950195, + "step": 959 + }, + { + "epoch": 13.335371179039301, + "grad_norm": 0.012328253127634525, + "learning_rate": 0.0025800251784308314, + "loss": 4.917585372924805, + "step": 960 + }, + { + "epoch": 13.34934497816594, + "grad_norm": 0.012000586837530136, + "learning_rate": 0.0025790958565205316, + "loss": 4.9137725830078125, + "step": 961 + }, + { + "epoch": 13.363318777292577, + "grad_norm": 0.009463997557759285, + "learning_rate": 0.002578165677287419, + "loss": 4.910362243652344, + "step": 962 + }, + { + "epoch": 13.377292576419213, + "grad_norm": 0.010591360740363598, + "learning_rate": 0.0025772346414809242, + "loss": 4.916032314300537, + "step": 963 + }, + { + "epoch": 13.391266375545852, + "grad_norm": 0.010429762303829193, + "learning_rate": 0.002576302749851164, + "loss": 4.934561729431152, + "step": 964 + }, + { + "epoch": 13.405240174672489, + "grad_norm": 0.12526583671569824, + "learning_rate": 0.0025753700031489486, + "loss": 4.897515773773193, + "step": 965 + }, + { + "epoch": 13.419213973799126, + "grad_norm": 0.01368892751634121, + "learning_rate": 0.0025744364021257743, + "loss": 4.94814920425415, + "step": 966 + }, + { + "epoch": 13.433187772925764, + "grad_norm": 0.016603967174887657, + "learning_rate": 0.0025735019475338254, + "loss": 4.8996758460998535, + "step": 967 + }, + { + "epoch": 13.447161572052401, + "grad_norm": 0.06110106781125069, + "learning_rate": 0.0025725666401259775, + "loss": 4.946794509887695, + "step": 968 + }, + { + "epoch": 13.46113537117904, + "grad_norm": 0.018292423337697983, + "learning_rate": 0.002571630480655789, + "loss": 4.958992958068848, + "step": 969 + }, + { + "epoch": 13.475109170305677, + "grad_norm": 0.01852245442569256, + "learning_rate": 0.0025706934698775073, + "loss": 4.942556381225586, + "step": 970 + }, + { + "epoch": 13.489082969432314, + "grad_norm": 0.0195649154484272, + "learning_rate": 0.0025697556085460646, + "loss": 4.895773887634277, + "step": 971 + }, + { + "epoch": 13.503056768558952, + "grad_norm": 0.017820244655013084, + "learning_rate": 0.00256881689741708, + "loss": 4.915250778198242, + "step": 972 + }, + { + "epoch": 13.51703056768559, + "grad_norm": 0.016814742237329483, + "learning_rate": 0.002567877337246855, + "loss": 4.9521260261535645, + "step": 973 + }, + { + "epoch": 13.531004366812226, + "grad_norm": 0.015319781377911568, + "learning_rate": 0.002566936928792377, + "loss": 4.899402141571045, + "step": 974 + }, + { + "epoch": 13.544978165938865, + "grad_norm": 0.01607401855289936, + "learning_rate": 0.0025659956728113146, + "loss": 4.909913063049316, + "step": 975 + }, + { + "epoch": 13.558951965065502, + "grad_norm": 0.015564955770969391, + "learning_rate": 0.002565053570062023, + "loss": 4.947719573974609, + "step": 976 + }, + { + "epoch": 13.57292576419214, + "grad_norm": 0.017502037808299065, + "learning_rate": 0.002564110621303536, + "loss": 4.95107364654541, + "step": 977 + }, + { + "epoch": 13.586899563318777, + "grad_norm": 0.014791073277592659, + "learning_rate": 0.002563166827295571, + "loss": 4.8995866775512695, + "step": 978 + }, + { + "epoch": 13.600873362445414, + "grad_norm": 0.012700124643743038, + "learning_rate": 0.002562222188798525, + "loss": 4.922100067138672, + "step": 979 + }, + { + "epoch": 13.614847161572053, + "grad_norm": 0.011235272511839867, + "learning_rate": 0.0025612767065734763, + "loss": 4.884456634521484, + "step": 980 + }, + { + "epoch": 13.62882096069869, + "grad_norm": 0.013009768910706043, + "learning_rate": 0.002560330381382183, + "loss": 4.869503021240234, + "step": 981 + }, + { + "epoch": 13.642794759825328, + "grad_norm": 0.0102638378739357, + "learning_rate": 0.0025593832139870834, + "loss": 4.934551239013672, + "step": 982 + }, + { + "epoch": 13.656768558951965, + "grad_norm": 0.011081870645284653, + "learning_rate": 0.0025584352051512913, + "loss": 4.937268257141113, + "step": 983 + }, + { + "epoch": 13.670742358078602, + "grad_norm": 0.010681218467652798, + "learning_rate": 0.002557486355638601, + "loss": 4.920559883117676, + "step": 984 + }, + { + "epoch": 13.68471615720524, + "grad_norm": 0.010962890461087227, + "learning_rate": 0.002556536666213484, + "loss": 4.936082363128662, + "step": 985 + }, + { + "epoch": 13.698689956331878, + "grad_norm": 0.010790379717946053, + "learning_rate": 0.0025555861376410868, + "loss": 4.9506449699401855, + "step": 986 + }, + { + "epoch": 13.712663755458514, + "grad_norm": 0.010659157298505306, + "learning_rate": 0.0025546347706872346, + "loss": 4.872883319854736, + "step": 987 + }, + { + "epoch": 13.726637554585153, + "grad_norm": 0.012962655164301395, + "learning_rate": 0.0025536825661184254, + "loss": 4.901210784912109, + "step": 988 + }, + { + "epoch": 13.74061135371179, + "grad_norm": 0.014810293912887573, + "learning_rate": 0.002552729524701833, + "loss": 4.895804405212402, + "step": 989 + }, + { + "epoch": 13.754585152838429, + "grad_norm": 0.010993832722306252, + "learning_rate": 0.0025517756472053068, + "loss": 4.891035079956055, + "step": 990 + }, + { + "epoch": 13.768558951965066, + "grad_norm": 0.013413915410637856, + "learning_rate": 0.002550820934397368, + "loss": 4.900317192077637, + "step": 991 + }, + { + "epoch": 13.782532751091702, + "grad_norm": 0.01580147072672844, + "learning_rate": 0.0025498653870472117, + "loss": 4.883669853210449, + "step": 992 + }, + { + "epoch": 13.796506550218341, + "grad_norm": 0.01864512264728546, + "learning_rate": 0.0025489090059247045, + "loss": 4.881938457489014, + "step": 993 + }, + { + "epoch": 13.810480349344978, + "grad_norm": 0.017185743898153305, + "learning_rate": 0.0025479517918003862, + "loss": 4.853325843811035, + "step": 994 + }, + { + "epoch": 13.824454148471617, + "grad_norm": 0.012045156210660934, + "learning_rate": 0.002546993745445466, + "loss": 4.9152326583862305, + "step": 995 + }, + { + "epoch": 13.838427947598253, + "grad_norm": 0.013420728035271168, + "learning_rate": 0.0025460348676318257, + "loss": 4.895130157470703, + "step": 996 + }, + { + "epoch": 13.85240174672489, + "grad_norm": 0.011561610735952854, + "learning_rate": 0.0025450751591320152, + "loss": 4.890257835388184, + "step": 997 + }, + { + "epoch": 13.866375545851529, + "grad_norm": 0.013164008036255836, + "learning_rate": 0.0025441146207192547, + "loss": 4.885972023010254, + "step": 998 + }, + { + "epoch": 13.880349344978166, + "grad_norm": 0.01120204571634531, + "learning_rate": 0.002543153253167432, + "loss": 4.882246971130371, + "step": 999 + }, + { + "epoch": 13.894323144104803, + "grad_norm": 0.011342627927660942, + "learning_rate": 0.002542191057251104, + "loss": 4.862927436828613, + "step": 1000 + }, + { + "epoch": 13.908296943231441, + "grad_norm": 0.010852213948965073, + "learning_rate": 0.0025412280337454942, + "loss": 4.879744529724121, + "step": 1001 + }, + { + "epoch": 13.922270742358078, + "grad_norm": 0.012851282022893429, + "learning_rate": 0.0025402641834264936, + "loss": 4.882238388061523, + "step": 1002 + }, + { + "epoch": 13.936244541484717, + "grad_norm": 0.011839655227959156, + "learning_rate": 0.002539299507070659, + "loss": 4.85658073425293, + "step": 1003 + }, + { + "epoch": 13.950218340611354, + "grad_norm": 0.011315262876451015, + "learning_rate": 0.002538334005455213, + "loss": 4.9234771728515625, + "step": 1004 + }, + { + "epoch": 13.96419213973799, + "grad_norm": 0.011661105789244175, + "learning_rate": 0.0025373676793580416, + "loss": 4.855696678161621, + "step": 1005 + }, + { + "epoch": 13.97816593886463, + "grad_norm": 0.009712521918118, + "learning_rate": 0.0025364005295576975, + "loss": 4.89942741394043, + "step": 1006 + }, + { + "epoch": 13.992139737991266, + "grad_norm": 0.011002928018569946, + "learning_rate": 0.0025354325568333956, + "loss": 4.871459007263184, + "step": 1007 + }, + { + "epoch": 14.0, + "grad_norm": 0.01509513333439827, + "learning_rate": 0.0025344637619650135, + "loss": 4.834778785705566, + "step": 1008 + }, + { + "epoch": 14.0, + "eval_loss": 5.031810283660889, + "eval_runtime": 51.9412, + "eval_samples_per_second": 47.015, + "eval_steps_per_second": 1.482, + "step": 1008 + }, + { + "epoch": 14.013973799126637, + "grad_norm": 0.010317740961909294, + "learning_rate": 0.0025334941457330925, + "loss": 4.8169708251953125, + "step": 1009 + }, + { + "epoch": 14.027947598253276, + "grad_norm": 0.00892715621739626, + "learning_rate": 0.0025325237089188343, + "loss": 4.852178573608398, + "step": 1010 + }, + { + "epoch": 14.041921397379912, + "grad_norm": 0.011608307249844074, + "learning_rate": 0.002531552452304103, + "loss": 4.790487289428711, + "step": 1011 + }, + { + "epoch": 14.055895196506551, + "grad_norm": 0.013100114651024342, + "learning_rate": 0.0025305803766714216, + "loss": 4.863536834716797, + "step": 1012 + }, + { + "epoch": 14.069868995633188, + "grad_norm": 0.016571057960391045, + "learning_rate": 0.002529607482803974, + "loss": 4.840987205505371, + "step": 1013 + }, + { + "epoch": 14.083842794759825, + "grad_norm": 0.014595622196793556, + "learning_rate": 0.002528633771485604, + "loss": 4.815027236938477, + "step": 1014 + }, + { + "epoch": 14.097816593886463, + "grad_norm": 0.014603931456804276, + "learning_rate": 0.002527659243500812, + "loss": 4.853024959564209, + "step": 1015 + }, + { + "epoch": 14.1117903930131, + "grad_norm": 0.019107041880488396, + "learning_rate": 0.0025266838996347595, + "loss": 4.829808235168457, + "step": 1016 + }, + { + "epoch": 14.125764192139737, + "grad_norm": 0.020292190834879875, + "learning_rate": 0.0025257077406732613, + "loss": 4.775232315063477, + "step": 1017 + }, + { + "epoch": 14.139737991266376, + "grad_norm": 0.014630074612796307, + "learning_rate": 0.0025247307674027926, + "loss": 4.826682090759277, + "step": 1018 + }, + { + "epoch": 14.153711790393013, + "grad_norm": 0.019507668912410736, + "learning_rate": 0.0025237529806104823, + "loss": 4.843222141265869, + "step": 1019 + }, + { + "epoch": 14.167685589519651, + "grad_norm": 0.023950330913066864, + "learning_rate": 0.002522774381084116, + "loss": 4.820512771606445, + "step": 1020 + }, + { + "epoch": 14.181659388646288, + "grad_norm": 0.022547820582985878, + "learning_rate": 0.002521794969612133, + "loss": 4.854910850524902, + "step": 1021 + }, + { + "epoch": 14.195633187772925, + "grad_norm": 0.017819277942180634, + "learning_rate": 0.0025208147469836283, + "loss": 4.825885772705078, + "step": 1022 + }, + { + "epoch": 14.209606986899564, + "grad_norm": 0.017989592626690865, + "learning_rate": 0.002519833713988349, + "loss": 4.8357014656066895, + "step": 1023 + }, + { + "epoch": 14.2235807860262, + "grad_norm": 0.013516640290617943, + "learning_rate": 0.0025188518714166956, + "loss": 4.83729362487793, + "step": 1024 + }, + { + "epoch": 14.237554585152838, + "grad_norm": 0.015956101939082146, + "learning_rate": 0.002517869220059721, + "loss": 4.807845592498779, + "step": 1025 + }, + { + "epoch": 14.251528384279476, + "grad_norm": 0.013967473991215229, + "learning_rate": 0.002516885760709131, + "loss": 4.828944206237793, + "step": 1026 + }, + { + "epoch": 14.265502183406113, + "grad_norm": 0.01270931027829647, + "learning_rate": 0.0025159014941572785, + "loss": 4.826594352722168, + "step": 1027 + }, + { + "epoch": 14.279475982532752, + "grad_norm": 0.012433716095983982, + "learning_rate": 0.002514916421197171, + "loss": 4.797939777374268, + "step": 1028 + }, + { + "epoch": 14.293449781659389, + "grad_norm": 0.011882808059453964, + "learning_rate": 0.0025139305426224636, + "loss": 4.78917932510376, + "step": 1029 + }, + { + "epoch": 14.307423580786025, + "grad_norm": 0.011193005368113518, + "learning_rate": 0.0025129438592274614, + "loss": 4.787964820861816, + "step": 1030 + }, + { + "epoch": 14.321397379912664, + "grad_norm": 0.011873407289385796, + "learning_rate": 0.002511956371807116, + "loss": 4.81754732131958, + "step": 1031 + }, + { + "epoch": 14.335371179039301, + "grad_norm": 0.011407418176531792, + "learning_rate": 0.002510968081157029, + "loss": 4.781639575958252, + "step": 1032 + }, + { + "epoch": 14.34934497816594, + "grad_norm": 0.009560304693877697, + "learning_rate": 0.002509978988073449, + "loss": 4.8163957595825195, + "step": 1033 + }, + { + "epoch": 14.363318777292577, + "grad_norm": 0.009969786740839481, + "learning_rate": 0.0025089890933532697, + "loss": 4.795045852661133, + "step": 1034 + }, + { + "epoch": 14.377292576419213, + "grad_norm": 0.010739767923951149, + "learning_rate": 0.0025079983977940312, + "loss": 4.847584247589111, + "step": 1035 + }, + { + "epoch": 14.391266375545852, + "grad_norm": 0.009999158792197704, + "learning_rate": 0.0025070069021939194, + "loss": 4.834909439086914, + "step": 1036 + }, + { + "epoch": 14.405240174672489, + "grad_norm": 0.009302482940256596, + "learning_rate": 0.002506014607351765, + "loss": 4.808745384216309, + "step": 1037 + }, + { + "epoch": 14.419213973799126, + "grad_norm": 0.010687744244933128, + "learning_rate": 0.0025050215140670403, + "loss": 4.789101600646973, + "step": 1038 + }, + { + "epoch": 14.433187772925764, + "grad_norm": 0.010930616408586502, + "learning_rate": 0.0025040276231398637, + "loss": 4.817464351654053, + "step": 1039 + }, + { + "epoch": 14.447161572052401, + "grad_norm": 0.013101065531373024, + "learning_rate": 0.0025030329353709956, + "loss": 4.827780723571777, + "step": 1040 + }, + { + "epoch": 14.46113537117904, + "grad_norm": 0.01223307941108942, + "learning_rate": 0.0025020374515618373, + "loss": 4.8367414474487305, + "step": 1041 + }, + { + "epoch": 14.475109170305677, + "grad_norm": 0.013091634958982468, + "learning_rate": 0.0025010411725144323, + "loss": 4.839454650878906, + "step": 1042 + }, + { + "epoch": 14.489082969432314, + "grad_norm": 0.012114843353629112, + "learning_rate": 0.0025000440990314643, + "loss": 4.832447052001953, + "step": 1043 + }, + { + "epoch": 14.503056768558952, + "grad_norm": 0.014664588496088982, + "learning_rate": 0.0024990462319162588, + "loss": 4.740812301635742, + "step": 1044 + }, + { + "epoch": 14.51703056768559, + "grad_norm": 0.015566283836960793, + "learning_rate": 0.002498047571972777, + "loss": 4.815671920776367, + "step": 1045 + }, + { + "epoch": 14.531004366812226, + "grad_norm": 0.012440434657037258, + "learning_rate": 0.002497048120005623, + "loss": 4.769775867462158, + "step": 1046 + }, + { + "epoch": 14.544978165938865, + "grad_norm": 0.01175499614328146, + "learning_rate": 0.0024960478768200362, + "loss": 4.8101348876953125, + "step": 1047 + }, + { + "epoch": 14.558951965065502, + "grad_norm": 0.012699508108198643, + "learning_rate": 0.002495046843221896, + "loss": 4.827281951904297, + "step": 1048 + }, + { + "epoch": 14.57292576419214, + "grad_norm": 0.012438303790986538, + "learning_rate": 0.002494045020017715, + "loss": 4.786289691925049, + "step": 1049 + }, + { + "epoch": 14.586899563318777, + "grad_norm": 0.01709924265742302, + "learning_rate": 0.002493042408014646, + "loss": 4.828742027282715, + "step": 1050 + }, + { + "epoch": 14.600873362445414, + "grad_norm": 0.022684535011649132, + "learning_rate": 0.0024920390080204733, + "loss": 4.822258949279785, + "step": 1051 + }, + { + "epoch": 14.614847161572053, + "grad_norm": 0.011984334327280521, + "learning_rate": 0.00249103482084362, + "loss": 4.792006969451904, + "step": 1052 + }, + { + "epoch": 14.62882096069869, + "grad_norm": 0.013941102661192417, + "learning_rate": 0.00249002984729314, + "loss": 4.7926177978515625, + "step": 1053 + }, + { + "epoch": 14.642794759825328, + "grad_norm": 0.0145634226500988, + "learning_rate": 0.002489024088178724, + "loss": 4.819155216217041, + "step": 1054 + }, + { + "epoch": 14.656768558951965, + "grad_norm": 0.013391612097620964, + "learning_rate": 0.0024880175443106932, + "loss": 4.777242660522461, + "step": 1055 + }, + { + "epoch": 14.670742358078602, + "grad_norm": 0.01764756441116333, + "learning_rate": 0.0024870102165000015, + "loss": 4.81627893447876, + "step": 1056 + }, + { + "epoch": 14.68471615720524, + "grad_norm": 0.015397441573441029, + "learning_rate": 0.002486002105558235, + "loss": 4.869495391845703, + "step": 1057 + }, + { + "epoch": 14.698689956331878, + "grad_norm": 0.01225620973855257, + "learning_rate": 0.0024849932122976105, + "loss": 4.784395217895508, + "step": 1058 + }, + { + "epoch": 14.712663755458514, + "grad_norm": 0.01534759160131216, + "learning_rate": 0.002483983537530975, + "loss": 4.813475608825684, + "step": 1059 + }, + { + "epoch": 14.726637554585153, + "grad_norm": 0.015057779848575592, + "learning_rate": 0.002482973082071805, + "loss": 4.8281569480896, + "step": 1060 + }, + { + "epoch": 14.74061135371179, + "grad_norm": 0.01243986189365387, + "learning_rate": 0.0024819618467342066, + "loss": 4.7866973876953125, + "step": 1061 + }, + { + "epoch": 14.754585152838429, + "grad_norm": 0.013326527550816536, + "learning_rate": 0.0024809498323329136, + "loss": 4.750214576721191, + "step": 1062 + }, + { + "epoch": 14.768558951965066, + "grad_norm": 0.013644695281982422, + "learning_rate": 0.002479937039683288, + "loss": 4.790144920349121, + "step": 1063 + }, + { + "epoch": 14.782532751091702, + "grad_norm": 0.01296213548630476, + "learning_rate": 0.002478923469601319, + "loss": 4.812170028686523, + "step": 1064 + }, + { + "epoch": 14.796506550218341, + "grad_norm": 0.011848201975226402, + "learning_rate": 0.0024779091229036204, + "loss": 4.798855781555176, + "step": 1065 + }, + { + "epoch": 14.810480349344978, + "grad_norm": 0.012765968218445778, + "learning_rate": 0.0024768940004074335, + "loss": 4.8272294998168945, + "step": 1066 + }, + { + "epoch": 14.824454148471617, + "grad_norm": 0.011700666509568691, + "learning_rate": 0.002475878102930625, + "loss": 4.762265205383301, + "step": 1067 + }, + { + "epoch": 14.838427947598253, + "grad_norm": 0.01018498931080103, + "learning_rate": 0.0024748614312916843, + "loss": 4.796628952026367, + "step": 1068 + }, + { + "epoch": 14.85240174672489, + "grad_norm": 0.010241681709885597, + "learning_rate": 0.002473843986309725, + "loss": 4.829319000244141, + "step": 1069 + }, + { + "epoch": 14.866375545851529, + "grad_norm": 0.009332080371677876, + "learning_rate": 0.0024728257688044855, + "loss": 4.783773899078369, + "step": 1070 + }, + { + "epoch": 14.880349344978166, + "grad_norm": 0.012860726565122604, + "learning_rate": 0.002471806779596324, + "loss": 4.802700519561768, + "step": 1071 + }, + { + "epoch": 14.894323144104803, + "grad_norm": 0.011156697757542133, + "learning_rate": 0.0024707870195062225, + "loss": 4.808038234710693, + "step": 1072 + }, + { + "epoch": 14.908296943231441, + "grad_norm": 0.010329720564186573, + "learning_rate": 0.002469766489355783, + "loss": 4.761645317077637, + "step": 1073 + }, + { + "epoch": 14.922270742358078, + "grad_norm": 0.011093895882368088, + "learning_rate": 0.0024687451899672278, + "loss": 4.795220851898193, + "step": 1074 + }, + { + "epoch": 14.936244541484717, + "grad_norm": 0.010740782134234905, + "learning_rate": 0.0024677231221634, + "loss": 4.79289436340332, + "step": 1075 + }, + { + "epoch": 14.950218340611354, + "grad_norm": 0.012272508814930916, + "learning_rate": 0.0024667002867677606, + "loss": 4.811873435974121, + "step": 1076 + }, + { + "epoch": 14.96419213973799, + "grad_norm": 0.008665879257023335, + "learning_rate": 0.0024656766846043905, + "loss": 4.744179725646973, + "step": 1077 + }, + { + "epoch": 14.97816593886463, + "grad_norm": 0.00933373998850584, + "learning_rate": 0.0024646523164979866, + "loss": 4.805453777313232, + "step": 1078 + }, + { + "epoch": 14.992139737991266, + "grad_norm": 0.010399337857961655, + "learning_rate": 0.002463627183273864, + "loss": 4.75568962097168, + "step": 1079 + }, + { + "epoch": 15.0, + "grad_norm": 0.013210986740887165, + "learning_rate": 0.0024626012857579542, + "loss": 4.797974586486816, + "step": 1080 + }, + { + "epoch": 15.0, + "eval_loss": 4.967026233673096, + "eval_runtime": 51.07, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 1080 + }, + { + "epoch": 15.013973799126637, + "grad_norm": 0.01731858029961586, + "learning_rate": 0.0024615746247768043, + "loss": 4.748838901519775, + "step": 1081 + }, + { + "epoch": 15.027947598253276, + "grad_norm": 0.018902689218521118, + "learning_rate": 0.0024605472011575763, + "loss": 4.717338562011719, + "step": 1082 + }, + { + "epoch": 15.041921397379912, + "grad_norm": 0.01734415628015995, + "learning_rate": 0.0024595190157280468, + "loss": 4.737737655639648, + "step": 1083 + }, + { + "epoch": 15.055895196506551, + "grad_norm": 0.01971321739256382, + "learning_rate": 0.002458490069316607, + "loss": 4.774558067321777, + "step": 1084 + }, + { + "epoch": 15.069868995633188, + "grad_norm": 0.020029844716191292, + "learning_rate": 0.002457460362752259, + "loss": 4.739579200744629, + "step": 1085 + }, + { + "epoch": 15.083842794759825, + "grad_norm": 0.020853109657764435, + "learning_rate": 0.0024564298968646203, + "loss": 4.791436672210693, + "step": 1086 + }, + { + "epoch": 15.097816593886463, + "grad_norm": 0.019159914925694466, + "learning_rate": 0.0024553986724839176, + "loss": 4.765683650970459, + "step": 1087 + }, + { + "epoch": 15.1117903930131, + "grad_norm": 0.0165473110973835, + "learning_rate": 0.0024543666904409903, + "loss": 4.732603073120117, + "step": 1088 + }, + { + "epoch": 15.125764192139737, + "grad_norm": 0.017426826059818268, + "learning_rate": 0.002453333951567287, + "loss": 4.7360639572143555, + "step": 1089 + }, + { + "epoch": 15.139737991266376, + "grad_norm": 0.014590734615921974, + "learning_rate": 0.002452300456694868, + "loss": 4.750031471252441, + "step": 1090 + }, + { + "epoch": 15.153711790393013, + "grad_norm": 0.011172222904860973, + "learning_rate": 0.0024512662066563996, + "loss": 4.740771293640137, + "step": 1091 + }, + { + "epoch": 15.167685589519651, + "grad_norm": 0.012294141575694084, + "learning_rate": 0.0024502312022851606, + "loss": 4.726214408874512, + "step": 1092 + }, + { + "epoch": 15.181659388646288, + "grad_norm": 0.01146282535046339, + "learning_rate": 0.002449195444415034, + "loss": 4.753599166870117, + "step": 1093 + }, + { + "epoch": 15.195633187772925, + "grad_norm": 0.011197247542440891, + "learning_rate": 0.0024481589338805113, + "loss": 4.751479148864746, + "step": 1094 + }, + { + "epoch": 15.209606986899564, + "grad_norm": 0.013252820819616318, + "learning_rate": 0.0024471216715166904, + "loss": 4.743409156799316, + "step": 1095 + }, + { + "epoch": 15.2235807860262, + "grad_norm": 0.012366211041808128, + "learning_rate": 0.0024460836581592744, + "loss": 4.731625556945801, + "step": 1096 + }, + { + "epoch": 15.237554585152838, + "grad_norm": 0.013906655833125114, + "learning_rate": 0.0024450448946445737, + "loss": 4.768153190612793, + "step": 1097 + }, + { + "epoch": 15.251528384279476, + "grad_norm": 0.012118754908442497, + "learning_rate": 0.002444005381809499, + "loss": 4.74198579788208, + "step": 1098 + }, + { + "epoch": 15.265502183406113, + "grad_norm": 0.01197550818324089, + "learning_rate": 0.002442965120491568, + "loss": 4.737289905548096, + "step": 1099 + }, + { + "epoch": 15.279475982532752, + "grad_norm": 0.010640372522175312, + "learning_rate": 0.002441924111528901, + "loss": 4.7472686767578125, + "step": 1100 + }, + { + "epoch": 15.293449781659389, + "grad_norm": 0.013553396798670292, + "learning_rate": 0.0024408823557602196, + "loss": 4.711215972900391, + "step": 1101 + }, + { + "epoch": 15.307423580786025, + "grad_norm": 0.01393711194396019, + "learning_rate": 0.002439839854024847, + "loss": 4.720060348510742, + "step": 1102 + }, + { + "epoch": 15.321397379912664, + "grad_norm": 0.01589774526655674, + "learning_rate": 0.002438796607162709, + "loss": 4.695697784423828, + "step": 1103 + }, + { + "epoch": 15.335371179039301, + "grad_norm": 0.0097908740863204, + "learning_rate": 0.0024377526160143303, + "loss": 4.729333877563477, + "step": 1104 + }, + { + "epoch": 15.34934497816594, + "grad_norm": 0.011745489202439785, + "learning_rate": 0.0024367078814208363, + "loss": 4.722957134246826, + "step": 1105 + }, + { + "epoch": 15.363318777292577, + "grad_norm": 0.009649662300944328, + "learning_rate": 0.0024356624042239498, + "loss": 4.727540493011475, + "step": 1106 + }, + { + "epoch": 15.377292576419213, + "grad_norm": 0.008429240435361862, + "learning_rate": 0.002434616185265993, + "loss": 4.705052375793457, + "step": 1107 + }, + { + "epoch": 15.391266375545852, + "grad_norm": 0.00984937883913517, + "learning_rate": 0.002433569225389887, + "loss": 4.721161365509033, + "step": 1108 + }, + { + "epoch": 15.405240174672489, + "grad_norm": 0.009237290360033512, + "learning_rate": 0.0024325215254391466, + "loss": 4.724277496337891, + "step": 1109 + }, + { + "epoch": 15.419213973799126, + "grad_norm": 0.008900545537471771, + "learning_rate": 0.002431473086257886, + "loss": 4.711522102355957, + "step": 1110 + }, + { + "epoch": 15.433187772925764, + "grad_norm": 0.010003026574850082, + "learning_rate": 0.002430423908690813, + "loss": 4.728816032409668, + "step": 1111 + }, + { + "epoch": 15.447161572052401, + "grad_norm": 0.011658324860036373, + "learning_rate": 0.0024293739935832316, + "loss": 4.716179847717285, + "step": 1112 + }, + { + "epoch": 15.46113537117904, + "grad_norm": 0.01312185451388359, + "learning_rate": 0.0024283233417810393, + "loss": 4.7172346115112305, + "step": 1113 + }, + { + "epoch": 15.475109170305677, + "grad_norm": 0.014306646771728992, + "learning_rate": 0.0024272719541307266, + "loss": 4.704054832458496, + "step": 1114 + }, + { + "epoch": 15.489082969432314, + "grad_norm": 0.009844634681940079, + "learning_rate": 0.0024262198314793783, + "loss": 4.72379207611084, + "step": 1115 + }, + { + "epoch": 15.503056768558952, + "grad_norm": 0.00903896801173687, + "learning_rate": 0.0024251669746746703, + "loss": 4.711025238037109, + "step": 1116 + }, + { + "epoch": 15.51703056768559, + "grad_norm": 0.010476433672010899, + "learning_rate": 0.0024241133845648703, + "loss": 4.711664199829102, + "step": 1117 + }, + { + "epoch": 15.531004366812226, + "grad_norm": 0.009128391742706299, + "learning_rate": 0.0024230590619988375, + "loss": 4.72395133972168, + "step": 1118 + }, + { + "epoch": 15.544978165938865, + "grad_norm": 0.01228832732886076, + "learning_rate": 0.0024220040078260195, + "loss": 4.747347831726074, + "step": 1119 + }, + { + "epoch": 15.558951965065502, + "grad_norm": 0.011916602961719036, + "learning_rate": 0.002420948222896455, + "loss": 4.696471691131592, + "step": 1120 + }, + { + "epoch": 15.57292576419214, + "grad_norm": 0.012877700850367546, + "learning_rate": 0.0024198917080607706, + "loss": 4.731119632720947, + "step": 1121 + }, + { + "epoch": 15.586899563318777, + "grad_norm": 0.012967764399945736, + "learning_rate": 0.0024188344641701814, + "loss": 4.688953876495361, + "step": 1122 + }, + { + "epoch": 15.600873362445414, + "grad_norm": 0.011993546038866043, + "learning_rate": 0.002417776492076489, + "loss": 4.695237159729004, + "step": 1123 + }, + { + "epoch": 15.614847161572053, + "grad_norm": 0.011188571341335773, + "learning_rate": 0.0024167177926320833, + "loss": 4.699850082397461, + "step": 1124 + }, + { + "epoch": 15.62882096069869, + "grad_norm": 0.010556348599493504, + "learning_rate": 0.0024156583666899395, + "loss": 4.703063011169434, + "step": 1125 + }, + { + "epoch": 15.642794759825328, + "grad_norm": 0.012023619376122952, + "learning_rate": 0.0024145982151036176, + "loss": 4.6918792724609375, + "step": 1126 + }, + { + "epoch": 15.656768558951965, + "grad_norm": 0.01266001258045435, + "learning_rate": 0.0024135373387272623, + "loss": 4.6653289794921875, + "step": 1127 + }, + { + "epoch": 15.670742358078602, + "grad_norm": 0.012152725830674171, + "learning_rate": 0.0024124757384156024, + "loss": 4.682897567749023, + "step": 1128 + }, + { + "epoch": 15.68471615720524, + "grad_norm": 0.014056944288313389, + "learning_rate": 0.002411413415023951, + "loss": 4.696136474609375, + "step": 1129 + }, + { + "epoch": 15.698689956331878, + "grad_norm": 0.016826556995511055, + "learning_rate": 0.002410350369408202, + "loss": 4.701657295227051, + "step": 1130 + }, + { + "epoch": 15.712663755458514, + "grad_norm": 0.015048800967633724, + "learning_rate": 0.0024092866024248323, + "loss": 4.691389083862305, + "step": 1131 + }, + { + "epoch": 15.726637554585153, + "grad_norm": 0.012512968853116035, + "learning_rate": 0.0024082221149308997, + "loss": 4.677886009216309, + "step": 1132 + }, + { + "epoch": 15.74061135371179, + "grad_norm": 0.01596425101161003, + "learning_rate": 0.0024071569077840425, + "loss": 4.718337535858154, + "step": 1133 + }, + { + "epoch": 15.754585152838429, + "grad_norm": 0.01640196517109871, + "learning_rate": 0.002406090981842479, + "loss": 4.697303295135498, + "step": 1134 + }, + { + "epoch": 15.768558951965066, + "grad_norm": 0.012347170151770115, + "learning_rate": 0.002405024337965007, + "loss": 4.696938991546631, + "step": 1135 + }, + { + "epoch": 15.782532751091702, + "grad_norm": 0.01213028747588396, + "learning_rate": 0.0024039569770110012, + "loss": 4.737306594848633, + "step": 1136 + }, + { + "epoch": 15.796506550218341, + "grad_norm": 0.01314542256295681, + "learning_rate": 0.002402888899840416, + "loss": 4.69948673248291, + "step": 1137 + }, + { + "epoch": 15.810480349344978, + "grad_norm": 0.013764189556241035, + "learning_rate": 0.0024018201073137805, + "loss": 4.722800254821777, + "step": 1138 + }, + { + "epoch": 15.824454148471617, + "grad_norm": 0.011147351004183292, + "learning_rate": 0.002400750600292203, + "loss": 4.694892883300781, + "step": 1139 + }, + { + "epoch": 15.838427947598253, + "grad_norm": 0.015437614172697067, + "learning_rate": 0.0023996803796373653, + "loss": 4.704752445220947, + "step": 1140 + }, + { + "epoch": 15.85240174672489, + "grad_norm": 0.016402097418904305, + "learning_rate": 0.0023986094462115255, + "loss": 4.67849063873291, + "step": 1141 + }, + { + "epoch": 15.866375545851529, + "grad_norm": 0.011600594967603683, + "learning_rate": 0.002397537800877514, + "loss": 4.689641952514648, + "step": 1142 + }, + { + "epoch": 15.880349344978166, + "grad_norm": 0.011864298023283482, + "learning_rate": 0.0023964654444987373, + "loss": 4.670377731323242, + "step": 1143 + }, + { + "epoch": 15.894323144104803, + "grad_norm": 0.01409872155636549, + "learning_rate": 0.002395392377939173, + "loss": 4.696022033691406, + "step": 1144 + }, + { + "epoch": 15.908296943231441, + "grad_norm": 0.012106554582715034, + "learning_rate": 0.0023943186020633717, + "loss": 4.705615043640137, + "step": 1145 + }, + { + "epoch": 15.922270742358078, + "grad_norm": 0.012310667894780636, + "learning_rate": 0.002393244117736455, + "loss": 4.702986717224121, + "step": 1146 + }, + { + "epoch": 15.936244541484717, + "grad_norm": 0.013501073233783245, + "learning_rate": 0.002392168925824116, + "loss": 4.6700215339660645, + "step": 1147 + }, + { + "epoch": 15.950218340611354, + "grad_norm": 0.013565126806497574, + "learning_rate": 0.0023910930271926166, + "loss": 4.699914455413818, + "step": 1148 + }, + { + "epoch": 15.96419213973799, + "grad_norm": 0.013421077281236649, + "learning_rate": 0.00239001642270879, + "loss": 4.713191986083984, + "step": 1149 + }, + { + "epoch": 15.97816593886463, + "grad_norm": 0.014408102259039879, + "learning_rate": 0.002388939113240035, + "loss": 4.655015468597412, + "step": 1150 + }, + { + "epoch": 15.992139737991266, + "grad_norm": 0.013502835296094418, + "learning_rate": 0.0023878610996543224, + "loss": 4.6142258644104, + "step": 1151 + }, + { + "epoch": 16.0, + "grad_norm": 0.014169570058584213, + "learning_rate": 0.002386782382820188, + "loss": 4.660243034362793, + "step": 1152 + }, + { + "epoch": 16.0, + "eval_loss": 4.885881423950195, + "eval_runtime": 52.0315, + "eval_samples_per_second": 46.933, + "eval_steps_per_second": 1.48, + "step": 1152 + }, + { + "epoch": 16.01397379912664, + "grad_norm": 0.014618083834648132, + "learning_rate": 0.002385702963606734, + "loss": 4.640046119689941, + "step": 1153 + }, + { + "epoch": 16.027947598253274, + "grad_norm": 0.01745622046291828, + "learning_rate": 0.0023846228428836278, + "loss": 4.641966819763184, + "step": 1154 + }, + { + "epoch": 16.041921397379912, + "grad_norm": 0.014995258301496506, + "learning_rate": 0.002383542021521106, + "loss": 4.60922908782959, + "step": 1155 + }, + { + "epoch": 16.05589519650655, + "grad_norm": 0.013278020545840263, + "learning_rate": 0.002382460500389964, + "loss": 4.652071952819824, + "step": 1156 + }, + { + "epoch": 16.069868995633186, + "grad_norm": 0.014685257337987423, + "learning_rate": 0.002381378280361566, + "loss": 4.672174453735352, + "step": 1157 + }, + { + "epoch": 16.083842794759825, + "grad_norm": 0.012737024575471878, + "learning_rate": 0.002380295362307836, + "loss": 4.612064361572266, + "step": 1158 + }, + { + "epoch": 16.097816593886463, + "grad_norm": 0.012992842122912407, + "learning_rate": 0.002379211747101261, + "loss": 4.611995697021484, + "step": 1159 + }, + { + "epoch": 16.111790393013102, + "grad_norm": 0.010210646316409111, + "learning_rate": 0.0023781274356148912, + "loss": 4.6055707931518555, + "step": 1160 + }, + { + "epoch": 16.125764192139737, + "grad_norm": 0.010265499353408813, + "learning_rate": 0.0023770424287223365, + "loss": 4.637089729309082, + "step": 1161 + }, + { + "epoch": 16.139737991266376, + "grad_norm": 0.010174009948968887, + "learning_rate": 0.002375956727297768, + "loss": 4.620571136474609, + "step": 1162 + }, + { + "epoch": 16.153711790393015, + "grad_norm": 0.009512446820735931, + "learning_rate": 0.0023748703322159143, + "loss": 4.644092559814453, + "step": 1163 + }, + { + "epoch": 16.16768558951965, + "grad_norm": 0.009849252179265022, + "learning_rate": 0.0023737832443520655, + "loss": 4.625055313110352, + "step": 1164 + }, + { + "epoch": 16.18165938864629, + "grad_norm": 0.010493188165128231, + "learning_rate": 0.0023726954645820674, + "loss": 4.644410610198975, + "step": 1165 + }, + { + "epoch": 16.195633187772927, + "grad_norm": 0.022181199863553047, + "learning_rate": 0.002371606993782326, + "loss": 4.593084335327148, + "step": 1166 + }, + { + "epoch": 16.209606986899562, + "grad_norm": 0.011906919069588184, + "learning_rate": 0.0023705178328298016, + "loss": 4.6452836990356445, + "step": 1167 + }, + { + "epoch": 16.2235807860262, + "grad_norm": 0.011874495074152946, + "learning_rate": 0.0023694279826020122, + "loss": 4.599322319030762, + "step": 1168 + }, + { + "epoch": 16.23755458515284, + "grad_norm": 0.00989708211272955, + "learning_rate": 0.0023683374439770295, + "loss": 4.607342720031738, + "step": 1169 + }, + { + "epoch": 16.251528384279474, + "grad_norm": 0.009256849065423012, + "learning_rate": 0.0023672462178334817, + "loss": 4.621583461761475, + "step": 1170 + }, + { + "epoch": 16.265502183406113, + "grad_norm": 0.013187132775783539, + "learning_rate": 0.0023661543050505496, + "loss": 4.614629745483398, + "step": 1171 + }, + { + "epoch": 16.27947598253275, + "grad_norm": 0.013505950570106506, + "learning_rate": 0.002365061706507968, + "loss": 4.629693984985352, + "step": 1172 + }, + { + "epoch": 16.29344978165939, + "grad_norm": 0.012032733298838139, + "learning_rate": 0.0023639684230860225, + "loss": 4.671209335327148, + "step": 1173 + }, + { + "epoch": 16.307423580786025, + "grad_norm": 0.01278337650001049, + "learning_rate": 0.002362874455665553, + "loss": 4.648874759674072, + "step": 1174 + }, + { + "epoch": 16.321397379912664, + "grad_norm": 0.01123690977692604, + "learning_rate": 0.0023617798051279496, + "loss": 4.611470699310303, + "step": 1175 + }, + { + "epoch": 16.335371179039303, + "grad_norm": 0.012564311735332012, + "learning_rate": 0.002360684472355151, + "loss": 4.659026145935059, + "step": 1176 + }, + { + "epoch": 16.349344978165938, + "grad_norm": 0.01206638477742672, + "learning_rate": 0.002359588458229649, + "loss": 4.642566204071045, + "step": 1177 + }, + { + "epoch": 16.363318777292577, + "grad_norm": 0.010300020687282085, + "learning_rate": 0.0023584917636344806, + "loss": 4.629585266113281, + "step": 1178 + }, + { + "epoch": 16.377292576419215, + "grad_norm": 0.0102265365421772, + "learning_rate": 0.0023573943894532336, + "loss": 4.609250068664551, + "step": 1179 + }, + { + "epoch": 16.39126637554585, + "grad_norm": 0.012440572492778301, + "learning_rate": 0.002356296336570042, + "loss": 4.636714458465576, + "step": 1180 + }, + { + "epoch": 16.40524017467249, + "grad_norm": 0.011830241419374943, + "learning_rate": 0.0023551976058695887, + "loss": 4.59100866317749, + "step": 1181 + }, + { + "epoch": 16.419213973799128, + "grad_norm": 0.011482699774205685, + "learning_rate": 0.0023540981982370997, + "loss": 4.620088577270508, + "step": 1182 + }, + { + "epoch": 16.433187772925763, + "grad_norm": 0.012950294651091099, + "learning_rate": 0.0023529981145583483, + "loss": 4.578657627105713, + "step": 1183 + }, + { + "epoch": 16.4471615720524, + "grad_norm": 0.012197710573673248, + "learning_rate": 0.002351897355719652, + "loss": 4.620813846588135, + "step": 1184 + }, + { + "epoch": 16.46113537117904, + "grad_norm": 0.014839230105280876, + "learning_rate": 0.002350795922607872, + "loss": 4.615731239318848, + "step": 1185 + }, + { + "epoch": 16.475109170305675, + "grad_norm": 0.016670791432261467, + "learning_rate": 0.002349693816110414, + "loss": 4.655855178833008, + "step": 1186 + }, + { + "epoch": 16.489082969432314, + "grad_norm": 0.014426100067794323, + "learning_rate": 0.0023485910371152247, + "loss": 4.660849571228027, + "step": 1187 + }, + { + "epoch": 16.503056768558952, + "grad_norm": 0.014040560461580753, + "learning_rate": 0.0023474875865107936, + "loss": 4.634917736053467, + "step": 1188 + }, + { + "epoch": 16.51703056768559, + "grad_norm": 0.018269391730427742, + "learning_rate": 0.0023463834651861496, + "loss": 4.611809253692627, + "step": 1189 + }, + { + "epoch": 16.531004366812226, + "grad_norm": 0.013855630531907082, + "learning_rate": 0.0023452786740308647, + "loss": 4.615225791931152, + "step": 1190 + }, + { + "epoch": 16.544978165938865, + "grad_norm": 0.011505477130413055, + "learning_rate": 0.0023441732139350487, + "loss": 4.62592887878418, + "step": 1191 + }, + { + "epoch": 16.558951965065503, + "grad_norm": 0.010166114196181297, + "learning_rate": 0.002343067085789351, + "loss": 4.6219024658203125, + "step": 1192 + }, + { + "epoch": 16.57292576419214, + "grad_norm": 0.009253021329641342, + "learning_rate": 0.0023419602904849596, + "loss": 4.658320426940918, + "step": 1193 + }, + { + "epoch": 16.586899563318777, + "grad_norm": 0.01027526892721653, + "learning_rate": 0.002340852828913598, + "loss": 4.589777946472168, + "step": 1194 + }, + { + "epoch": 16.600873362445416, + "grad_norm": 0.008855022490024567, + "learning_rate": 0.002339744701967529, + "loss": 4.588204383850098, + "step": 1195 + }, + { + "epoch": 16.61484716157205, + "grad_norm": 0.00965463649481535, + "learning_rate": 0.002338635910539551, + "loss": 4.648177146911621, + "step": 1196 + }, + { + "epoch": 16.62882096069869, + "grad_norm": 0.009073720313608646, + "learning_rate": 0.0023375264555229956, + "loss": 4.585582733154297, + "step": 1197 + }, + { + "epoch": 16.64279475982533, + "grad_norm": 0.010105383582413197, + "learning_rate": 0.0023364163378117317, + "loss": 4.612174987792969, + "step": 1198 + }, + { + "epoch": 16.656768558951963, + "grad_norm": 0.015307556837797165, + "learning_rate": 0.002335305558300161, + "loss": 4.591984748840332, + "step": 1199 + }, + { + "epoch": 16.670742358078602, + "grad_norm": 0.016716254875063896, + "learning_rate": 0.002334194117883219, + "loss": 4.659285545349121, + "step": 1200 + }, + { + "epoch": 16.68471615720524, + "grad_norm": 0.013281598687171936, + "learning_rate": 0.0023330820174563725, + "loss": 4.645829200744629, + "step": 1201 + }, + { + "epoch": 16.69868995633188, + "grad_norm": 0.014416513964533806, + "learning_rate": 0.0023319692579156204, + "loss": 4.584893226623535, + "step": 1202 + }, + { + "epoch": 16.712663755458514, + "grad_norm": 0.016326071694493294, + "learning_rate": 0.002330855840157494, + "loss": 4.633223533630371, + "step": 1203 + }, + { + "epoch": 16.726637554585153, + "grad_norm": 0.012851824052631855, + "learning_rate": 0.002329741765079053, + "loss": 4.580986499786377, + "step": 1204 + }, + { + "epoch": 16.74061135371179, + "grad_norm": 0.015444506891071796, + "learning_rate": 0.0023286270335778876, + "loss": 4.615891456604004, + "step": 1205 + }, + { + "epoch": 16.754585152838427, + "grad_norm": 0.01391979493200779, + "learning_rate": 0.0023275116465521174, + "loss": 4.628403186798096, + "step": 1206 + }, + { + "epoch": 16.768558951965066, + "grad_norm": 0.012138686142861843, + "learning_rate": 0.0023263956049003895, + "loss": 4.6259942054748535, + "step": 1207 + }, + { + "epoch": 16.782532751091704, + "grad_norm": 0.013741672039031982, + "learning_rate": 0.0023252789095218785, + "loss": 4.6056742668151855, + "step": 1208 + }, + { + "epoch": 16.79650655021834, + "grad_norm": 0.01103319600224495, + "learning_rate": 0.002324161561316285, + "loss": 4.621670246124268, + "step": 1209 + }, + { + "epoch": 16.810480349344978, + "grad_norm": 0.010333316400647163, + "learning_rate": 0.0023230435611838367, + "loss": 4.620484352111816, + "step": 1210 + }, + { + "epoch": 16.824454148471617, + "grad_norm": 0.011063606478273869, + "learning_rate": 0.0023219249100252864, + "loss": 4.64952278137207, + "step": 1211 + }, + { + "epoch": 16.83842794759825, + "grad_norm": 0.009984857402741909, + "learning_rate": 0.0023208056087419106, + "loss": 4.572571754455566, + "step": 1212 + }, + { + "epoch": 16.85240174672489, + "grad_norm": 0.009986856952309608, + "learning_rate": 0.002319685658235511, + "loss": 4.570366382598877, + "step": 1213 + }, + { + "epoch": 16.86637554585153, + "grad_norm": 0.010905452072620392, + "learning_rate": 0.0023185650594084103, + "loss": 4.603614807128906, + "step": 1214 + }, + { + "epoch": 16.880349344978168, + "grad_norm": 0.012800061143934727, + "learning_rate": 0.002317443813163456, + "loss": 4.562705039978027, + "step": 1215 + }, + { + "epoch": 16.894323144104803, + "grad_norm": 0.01252608560025692, + "learning_rate": 0.002316321920404015, + "loss": 4.599016189575195, + "step": 1216 + }, + { + "epoch": 16.90829694323144, + "grad_norm": 0.011305592954158783, + "learning_rate": 0.002315199382033976, + "loss": 4.644617557525635, + "step": 1217 + }, + { + "epoch": 16.92227074235808, + "grad_norm": 0.010594413615763187, + "learning_rate": 0.002314076198957749, + "loss": 4.569808006286621, + "step": 1218 + }, + { + "epoch": 16.936244541484715, + "grad_norm": 0.009859138168394566, + "learning_rate": 0.002312952372080261, + "loss": 4.5890398025512695, + "step": 1219 + }, + { + "epoch": 16.950218340611354, + "grad_norm": 0.010982917621731758, + "learning_rate": 0.00231182790230696, + "loss": 4.606348991394043, + "step": 1220 + }, + { + "epoch": 16.964192139737992, + "grad_norm": 0.010252755135297775, + "learning_rate": 0.00231070279054381, + "loss": 4.58942174911499, + "step": 1221 + }, + { + "epoch": 16.978165938864628, + "grad_norm": 0.01110257301479578, + "learning_rate": 0.0023095770376972947, + "loss": 4.599823474884033, + "step": 1222 + }, + { + "epoch": 16.992139737991266, + "grad_norm": 0.01003815233707428, + "learning_rate": 0.0023084506446744103, + "loss": 4.569343090057373, + "step": 1223 + }, + { + "epoch": 17.0, + "grad_norm": 0.010517258197069168, + "learning_rate": 0.0023073236123826745, + "loss": 4.646301746368408, + "step": 1224 + }, + { + "epoch": 17.0, + "eval_loss": 4.804611682891846, + "eval_runtime": 51.6279, + "eval_samples_per_second": 47.3, + "eval_steps_per_second": 1.491, + "step": 1224 + }, + { + "epoch": 17.01397379912664, + "grad_norm": 0.011439066380262375, + "learning_rate": 0.0023061959417301145, + "loss": 4.563335418701172, + "step": 1225 + }, + { + "epoch": 17.027947598253274, + "grad_norm": 0.01073157787322998, + "learning_rate": 0.002305067633625275, + "loss": 4.540966987609863, + "step": 1226 + }, + { + "epoch": 17.041921397379912, + "grad_norm": 0.011391228064894676, + "learning_rate": 0.0023039386889772136, + "loss": 4.555583953857422, + "step": 1227 + }, + { + "epoch": 17.05589519650655, + "grad_norm": 0.012814595364034176, + "learning_rate": 0.0023028091086955, + "loss": 4.517425537109375, + "step": 1228 + }, + { + "epoch": 17.069868995633186, + "grad_norm": 0.012994988821446896, + "learning_rate": 0.002301678893690218, + "loss": 4.538477897644043, + "step": 1229 + }, + { + "epoch": 17.083842794759825, + "grad_norm": 0.01460991520434618, + "learning_rate": 0.00230054804487196, + "loss": 4.571145057678223, + "step": 1230 + }, + { + "epoch": 17.097816593886463, + "grad_norm": 0.013036075979471207, + "learning_rate": 0.0022994165631518306, + "loss": 4.519233703613281, + "step": 1231 + }, + { + "epoch": 17.111790393013102, + "grad_norm": 0.011917057447135448, + "learning_rate": 0.0022982844494414444, + "loss": 4.5071587562561035, + "step": 1232 + }, + { + "epoch": 17.125764192139737, + "grad_norm": 0.013928866013884544, + "learning_rate": 0.0022971517046529257, + "loss": 4.531422138214111, + "step": 1233 + }, + { + "epoch": 17.139737991266376, + "grad_norm": 0.01407038513571024, + "learning_rate": 0.0022960183296989053, + "loss": 4.511205196380615, + "step": 1234 + }, + { + "epoch": 17.153711790393015, + "grad_norm": 0.019396567717194557, + "learning_rate": 0.0022948843254925245, + "loss": 4.530007362365723, + "step": 1235 + }, + { + "epoch": 17.16768558951965, + "grad_norm": 0.014425835572183132, + "learning_rate": 0.002293749692947428, + "loss": 4.530138969421387, + "step": 1236 + }, + { + "epoch": 17.18165938864629, + "grad_norm": 0.01600247249007225, + "learning_rate": 0.00229261443297777, + "loss": 4.5643391609191895, + "step": 1237 + }, + { + "epoch": 17.195633187772927, + "grad_norm": 0.018238037824630737, + "learning_rate": 0.0022914785464982095, + "loss": 4.540779113769531, + "step": 1238 + }, + { + "epoch": 17.209606986899562, + "grad_norm": 0.015427535399794579, + "learning_rate": 0.0022903420344239085, + "loss": 4.506287097930908, + "step": 1239 + }, + { + "epoch": 17.2235807860262, + "grad_norm": 0.016816625371575356, + "learning_rate": 0.0022892048976705347, + "loss": 4.537919521331787, + "step": 1240 + }, + { + "epoch": 17.23755458515284, + "grad_norm": 0.01999717950820923, + "learning_rate": 0.002288067137154259, + "loss": 4.558426380157471, + "step": 1241 + }, + { + "epoch": 17.251528384279474, + "grad_norm": 0.020884918048977852, + "learning_rate": 0.002286928753791753, + "loss": 4.592915058135986, + "step": 1242 + }, + { + "epoch": 17.265502183406113, + "grad_norm": 0.01620141603052616, + "learning_rate": 0.002285789748500194, + "loss": 4.553356170654297, + "step": 1243 + }, + { + "epoch": 17.27947598253275, + "grad_norm": 0.018495604395866394, + "learning_rate": 0.0022846501221972567, + "loss": 4.573139190673828, + "step": 1244 + }, + { + "epoch": 17.29344978165939, + "grad_norm": 0.0156320258975029, + "learning_rate": 0.0022835098758011174, + "loss": 4.585921764373779, + "step": 1245 + }, + { + "epoch": 17.307423580786025, + "grad_norm": 0.014515317976474762, + "learning_rate": 0.002282369010230452, + "loss": 4.5532002449035645, + "step": 1246 + }, + { + "epoch": 17.321397379912664, + "grad_norm": 0.014826556667685509, + "learning_rate": 0.0022812275264044355, + "loss": 4.560998916625977, + "step": 1247 + }, + { + "epoch": 17.335371179039303, + "grad_norm": 0.013240831904113293, + "learning_rate": 0.00228008542524274, + "loss": 4.531787872314453, + "step": 1248 + }, + { + "epoch": 17.349344978165938, + "grad_norm": 0.013814719393849373, + "learning_rate": 0.0022789427076655372, + "loss": 4.611872673034668, + "step": 1249 + }, + { + "epoch": 17.363318777292577, + "grad_norm": 0.012246561236679554, + "learning_rate": 0.0022777993745934934, + "loss": 4.5565385818481445, + "step": 1250 + }, + { + "epoch": 17.377292576419215, + "grad_norm": 0.01206218171864748, + "learning_rate": 0.002276655426947771, + "loss": 4.570852279663086, + "step": 1251 + }, + { + "epoch": 17.39126637554585, + "grad_norm": 0.01007329672574997, + "learning_rate": 0.0022755108656500275, + "loss": 4.553400039672852, + "step": 1252 + }, + { + "epoch": 17.40524017467249, + "grad_norm": 0.010361720807850361, + "learning_rate": 0.0022743656916224173, + "loss": 4.527509689331055, + "step": 1253 + }, + { + "epoch": 17.419213973799128, + "grad_norm": 0.009456256404519081, + "learning_rate": 0.0022732199057875846, + "loss": 4.520236015319824, + "step": 1254 + }, + { + "epoch": 17.433187772925763, + "grad_norm": 0.009310049936175346, + "learning_rate": 0.002272073509068669, + "loss": 4.518343925476074, + "step": 1255 + }, + { + "epoch": 17.4471615720524, + "grad_norm": 0.00927449855953455, + "learning_rate": 0.0022709265023893015, + "loss": 4.559690475463867, + "step": 1256 + }, + { + "epoch": 17.46113537117904, + "grad_norm": 0.008658594451844692, + "learning_rate": 0.002269778886673605, + "loss": 4.513876914978027, + "step": 1257 + }, + { + "epoch": 17.475109170305675, + "grad_norm": 0.008928220719099045, + "learning_rate": 0.0022686306628461918, + "loss": 4.545050621032715, + "step": 1258 + }, + { + "epoch": 17.489082969432314, + "grad_norm": 0.008046358823776245, + "learning_rate": 0.002267481831832166, + "loss": 4.54866361618042, + "step": 1259 + }, + { + "epoch": 17.503056768558952, + "grad_norm": 0.008318924345076084, + "learning_rate": 0.00226633239455712, + "loss": 4.546968936920166, + "step": 1260 + }, + { + "epoch": 17.51703056768559, + "grad_norm": 0.008808869868516922, + "learning_rate": 0.0022651823519471343, + "loss": 4.50877571105957, + "step": 1261 + }, + { + "epoch": 17.531004366812226, + "grad_norm": 0.008767383173108101, + "learning_rate": 0.0022640317049287774, + "loss": 4.57666540145874, + "step": 1262 + }, + { + "epoch": 17.544978165938865, + "grad_norm": 0.009351354092359543, + "learning_rate": 0.0022628804544291047, + "loss": 4.511702537536621, + "step": 1263 + }, + { + "epoch": 17.558951965065503, + "grad_norm": 0.010137057863175869, + "learning_rate": 0.002261728601375659, + "loss": 4.539676189422607, + "step": 1264 + }, + { + "epoch": 17.57292576419214, + "grad_norm": 0.00882046204060316, + "learning_rate": 0.002260576146696466, + "loss": 4.533196449279785, + "step": 1265 + }, + { + "epoch": 17.586899563318777, + "grad_norm": 0.009354782290756702, + "learning_rate": 0.0022594230913200388, + "loss": 4.515488624572754, + "step": 1266 + }, + { + "epoch": 17.600873362445416, + "grad_norm": 0.009241201914846897, + "learning_rate": 0.002258269436175372, + "loss": 4.542706489562988, + "step": 1267 + }, + { + "epoch": 17.61484716157205, + "grad_norm": 0.009252829477190971, + "learning_rate": 0.002257115182191947, + "loss": 4.592660903930664, + "step": 1268 + }, + { + "epoch": 17.62882096069869, + "grad_norm": 0.011243131943047047, + "learning_rate": 0.0022559603302997225, + "loss": 4.527804851531982, + "step": 1269 + }, + { + "epoch": 17.64279475982533, + "grad_norm": 0.011032823473215103, + "learning_rate": 0.0022548048814291437, + "loss": 4.540122032165527, + "step": 1270 + }, + { + "epoch": 17.656768558951963, + "grad_norm": 0.010495275259017944, + "learning_rate": 0.0022536488365111347, + "loss": 4.532835006713867, + "step": 1271 + }, + { + "epoch": 17.670742358078602, + "grad_norm": 0.009992090053856373, + "learning_rate": 0.0022524921964770994, + "loss": 4.521217346191406, + "step": 1272 + }, + { + "epoch": 17.68471615720524, + "grad_norm": 0.01141198631376028, + "learning_rate": 0.0022513349622589225, + "loss": 4.519709587097168, + "step": 1273 + }, + { + "epoch": 17.69868995633188, + "grad_norm": 0.013087889179587364, + "learning_rate": 0.002250177134788966, + "loss": 4.547289848327637, + "step": 1274 + }, + { + "epoch": 17.712663755458514, + "grad_norm": 0.01328477542847395, + "learning_rate": 0.0022490187150000706, + "loss": 4.527865409851074, + "step": 1275 + }, + { + "epoch": 17.726637554585153, + "grad_norm": 0.01033794041723013, + "learning_rate": 0.002247859703825555, + "loss": 4.54740047454834, + "step": 1276 + }, + { + "epoch": 17.74061135371179, + "grad_norm": 0.010522300377488136, + "learning_rate": 0.002246700102199213, + "loss": 4.538840293884277, + "step": 1277 + }, + { + "epoch": 17.754585152838427, + "grad_norm": 0.012151090428233147, + "learning_rate": 0.0022455399110553145, + "loss": 4.536865234375, + "step": 1278 + }, + { + "epoch": 17.768558951965066, + "grad_norm": 0.010553102008998394, + "learning_rate": 0.0022443791313286054, + "loss": 4.5348334312438965, + "step": 1279 + }, + { + "epoch": 17.782532751091704, + "grad_norm": 0.011433709412813187, + "learning_rate": 0.0022432177639543035, + "loss": 4.510526657104492, + "step": 1280 + }, + { + "epoch": 17.79650655021834, + "grad_norm": 0.01189691387116909, + "learning_rate": 0.002242055809868103, + "loss": 4.530807018280029, + "step": 1281 + }, + { + "epoch": 17.810480349344978, + "grad_norm": 0.010031254962086678, + "learning_rate": 0.0022408932700061683, + "loss": 4.496850967407227, + "step": 1282 + }, + { + "epoch": 17.824454148471617, + "grad_norm": 0.010129331611096859, + "learning_rate": 0.002239730145305138, + "loss": 4.551287651062012, + "step": 1283 + }, + { + "epoch": 17.83842794759825, + "grad_norm": 0.010430586524307728, + "learning_rate": 0.0022385664367021185, + "loss": 4.542672157287598, + "step": 1284 + }, + { + "epoch": 17.85240174672489, + "grad_norm": 0.011774476617574692, + "learning_rate": 0.002237402145134691, + "loss": 4.537034034729004, + "step": 1285 + }, + { + "epoch": 17.86637554585153, + "grad_norm": 0.01184659544378519, + "learning_rate": 0.002236237271540903, + "loss": 4.521193504333496, + "step": 1286 + }, + { + "epoch": 17.880349344978168, + "grad_norm": 0.011327247135341167, + "learning_rate": 0.0022350718168592724, + "loss": 4.49426794052124, + "step": 1287 + }, + { + "epoch": 17.894323144104803, + "grad_norm": 0.010767918080091476, + "learning_rate": 0.0022339057820287855, + "loss": 4.531225204467773, + "step": 1288 + }, + { + "epoch": 17.90829694323144, + "grad_norm": 0.011195424944162369, + "learning_rate": 0.002232739167988894, + "loss": 4.492671012878418, + "step": 1289 + }, + { + "epoch": 17.92227074235808, + "grad_norm": 0.010394908487796783, + "learning_rate": 0.0022315719756795197, + "loss": 4.523616790771484, + "step": 1290 + }, + { + "epoch": 17.936244541484715, + "grad_norm": 0.010277998633682728, + "learning_rate": 0.002230404206041047, + "loss": 4.487832069396973, + "step": 1291 + }, + { + "epoch": 17.950218340611354, + "grad_norm": 0.010955783538520336, + "learning_rate": 0.0022292358600143275, + "loss": 4.526865005493164, + "step": 1292 + }, + { + "epoch": 17.964192139737992, + "grad_norm": 0.010611076839268208, + "learning_rate": 0.002228066938540676, + "loss": 4.498682022094727, + "step": 1293 + }, + { + "epoch": 17.978165938864628, + "grad_norm": 0.010008828714489937, + "learning_rate": 0.0022268974425618724, + "loss": 4.524019241333008, + "step": 1294 + }, + { + "epoch": 17.992139737991266, + "grad_norm": 0.008973540738224983, + "learning_rate": 0.002225727373020157, + "loss": 4.512996673583984, + "step": 1295 + }, + { + "epoch": 18.0, + "grad_norm": 0.011754963546991348, + "learning_rate": 0.002224556730858235, + "loss": 4.496376991271973, + "step": 1296 + }, + { + "epoch": 18.0, + "eval_loss": 4.744190216064453, + "eval_runtime": 51.6572, + "eval_samples_per_second": 47.273, + "eval_steps_per_second": 1.491, + "step": 1296 + }, + { + "epoch": 18.01397379912664, + "grad_norm": 0.012094628997147083, + "learning_rate": 0.0022233855170192703, + "loss": 4.443447113037109, + "step": 1297 + }, + { + "epoch": 18.027947598253274, + "grad_norm": 0.014347990974783897, + "learning_rate": 0.00222221373244689, + "loss": 4.485690116882324, + "step": 1298 + }, + { + "epoch": 18.041921397379912, + "grad_norm": 0.021286984905600548, + "learning_rate": 0.0022210413780851786, + "loss": 4.487505912780762, + "step": 1299 + }, + { + "epoch": 18.05589519650655, + "grad_norm": 0.020273910835385323, + "learning_rate": 0.0022198684548786827, + "loss": 4.473119735717773, + "step": 1300 + }, + { + "epoch": 18.069868995633186, + "grad_norm": 0.017642971128225327, + "learning_rate": 0.002218694963772403, + "loss": 4.440615653991699, + "step": 1301 + }, + { + "epoch": 18.083842794759825, + "grad_norm": 0.02075147069990635, + "learning_rate": 0.0022175209057118017, + "loss": 4.527026176452637, + "step": 1302 + }, + { + "epoch": 18.097816593886463, + "grad_norm": 0.016479508951306343, + "learning_rate": 0.002216346281642796, + "loss": 4.508321762084961, + "step": 1303 + }, + { + "epoch": 18.111790393013102, + "grad_norm": 0.017978593707084656, + "learning_rate": 0.002215171092511759, + "loss": 4.500985145568848, + "step": 1304 + }, + { + "epoch": 18.125764192139737, + "grad_norm": 0.01739897020161152, + "learning_rate": 0.002213995339265519, + "loss": 4.474225044250488, + "step": 1305 + }, + { + "epoch": 18.139737991266376, + "grad_norm": 0.015556776896119118, + "learning_rate": 0.0022128190228513597, + "loss": 4.444849014282227, + "step": 1306 + }, + { + "epoch": 18.153711790393015, + "grad_norm": 0.016166578978300095, + "learning_rate": 0.002211642144217018, + "loss": 4.5026726722717285, + "step": 1307 + }, + { + "epoch": 18.16768558951965, + "grad_norm": 0.01534261740744114, + "learning_rate": 0.0022104647043106832, + "loss": 4.484527587890625, + "step": 1308 + }, + { + "epoch": 18.18165938864629, + "grad_norm": 0.013517106883227825, + "learning_rate": 0.002209286704080998, + "loss": 4.453907012939453, + "step": 1309 + }, + { + "epoch": 18.195633187772927, + "grad_norm": 0.012562787160277367, + "learning_rate": 0.0022081081444770556, + "loss": 4.471528053283691, + "step": 1310 + }, + { + "epoch": 18.209606986899562, + "grad_norm": 0.009947682730853558, + "learning_rate": 0.0022069290264483997, + "loss": 4.504153728485107, + "step": 1311 + }, + { + "epoch": 18.2235807860262, + "grad_norm": 0.01169571466743946, + "learning_rate": 0.002205749350945026, + "loss": 4.447345733642578, + "step": 1312 + }, + { + "epoch": 18.23755458515284, + "grad_norm": 0.011394352652132511, + "learning_rate": 0.0022045691189173755, + "loss": 4.447491645812988, + "step": 1313 + }, + { + "epoch": 18.251528384279474, + "grad_norm": 0.011168593540787697, + "learning_rate": 0.0022033883313163417, + "loss": 4.441639423370361, + "step": 1314 + }, + { + "epoch": 18.265502183406113, + "grad_norm": 0.009503341279923916, + "learning_rate": 0.002202206989093263, + "loss": 4.461601257324219, + "step": 1315 + }, + { + "epoch": 18.27947598253275, + "grad_norm": 0.0112199941650033, + "learning_rate": 0.002201025093199926, + "loss": 4.444095134735107, + "step": 1316 + }, + { + "epoch": 18.29344978165939, + "grad_norm": 0.010505333542823792, + "learning_rate": 0.002199842644588562, + "loss": 4.460987091064453, + "step": 1317 + }, + { + "epoch": 18.307423580786025, + "grad_norm": 0.009494813159108162, + "learning_rate": 0.0021986596442118496, + "loss": 4.460158348083496, + "step": 1318 + }, + { + "epoch": 18.321397379912664, + "grad_norm": 0.01063452661037445, + "learning_rate": 0.00219747609302291, + "loss": 4.489625453948975, + "step": 1319 + }, + { + "epoch": 18.335371179039303, + "grad_norm": 0.012194888666272163, + "learning_rate": 0.0021962919919753087, + "loss": 4.441605567932129, + "step": 1320 + }, + { + "epoch": 18.349344978165938, + "grad_norm": 0.010613647289574146, + "learning_rate": 0.0021951073420230557, + "loss": 4.475821018218994, + "step": 1321 + }, + { + "epoch": 18.363318777292577, + "grad_norm": 0.009643700905144215, + "learning_rate": 0.002193922144120601, + "loss": 4.467228889465332, + "step": 1322 + }, + { + "epoch": 18.377292576419215, + "grad_norm": 0.00912613607943058, + "learning_rate": 0.0021927363992228373, + "loss": 4.455435752868652, + "step": 1323 + }, + { + "epoch": 18.39126637554585, + "grad_norm": 0.009371735155582428, + "learning_rate": 0.0021915501082850988, + "loss": 4.479915142059326, + "step": 1324 + }, + { + "epoch": 18.40524017467249, + "grad_norm": 0.008617875166237354, + "learning_rate": 0.002190363272263158, + "loss": 4.415642738342285, + "step": 1325 + }, + { + "epoch": 18.419213973799128, + "grad_norm": 0.009571383707225323, + "learning_rate": 0.002189175892113227, + "loss": 4.439854621887207, + "step": 1326 + }, + { + "epoch": 18.433187772925763, + "grad_norm": 0.00782635249197483, + "learning_rate": 0.002187987968791957, + "loss": 4.484311103820801, + "step": 1327 + }, + { + "epoch": 18.4471615720524, + "grad_norm": 0.008408057503402233, + "learning_rate": 0.0021867995032564358, + "loss": 4.46577262878418, + "step": 1328 + }, + { + "epoch": 18.46113537117904, + "grad_norm": 0.009641910903155804, + "learning_rate": 0.00218561049646419, + "loss": 4.437408447265625, + "step": 1329 + }, + { + "epoch": 18.475109170305675, + "grad_norm": 0.008498182520270348, + "learning_rate": 0.0021844209493731784, + "loss": 4.442599296569824, + "step": 1330 + }, + { + "epoch": 18.489082969432314, + "grad_norm": 0.009289798326790333, + "learning_rate": 0.0021832308629418, + "loss": 4.517641067504883, + "step": 1331 + }, + { + "epoch": 18.503056768558952, + "grad_norm": 0.009360183961689472, + "learning_rate": 0.0021820402381288848, + "loss": 4.474143981933594, + "step": 1332 + }, + { + "epoch": 18.51703056768559, + "grad_norm": 0.008764738216996193, + "learning_rate": 0.0021808490758936983, + "loss": 4.4376702308654785, + "step": 1333 + }, + { + "epoch": 18.531004366812226, + "grad_norm": 0.009673701599240303, + "learning_rate": 0.0021796573771959375, + "loss": 4.435220718383789, + "step": 1334 + }, + { + "epoch": 18.544978165938865, + "grad_norm": 0.010852573439478874, + "learning_rate": 0.0021784651429957334, + "loss": 4.465187072753906, + "step": 1335 + }, + { + "epoch": 18.558951965065503, + "grad_norm": 0.011292314156889915, + "learning_rate": 0.0021772723742536464, + "loss": 4.458176136016846, + "step": 1336 + }, + { + "epoch": 18.57292576419214, + "grad_norm": 0.01071334257721901, + "learning_rate": 0.0021760790719306702, + "loss": 4.425498962402344, + "step": 1337 + }, + { + "epoch": 18.586899563318777, + "grad_norm": 0.01505239401012659, + "learning_rate": 0.0021748852369882265, + "loss": 4.463129043579102, + "step": 1338 + }, + { + "epoch": 18.600873362445416, + "grad_norm": 0.011774689890444279, + "learning_rate": 0.0021736908703881663, + "loss": 4.462577819824219, + "step": 1339 + }, + { + "epoch": 18.61484716157205, + "grad_norm": 0.011241118423640728, + "learning_rate": 0.002172495973092769, + "loss": 4.452583312988281, + "step": 1340 + }, + { + "epoch": 18.62882096069869, + "grad_norm": 0.01072910986840725, + "learning_rate": 0.002171300546064742, + "loss": 4.466248035430908, + "step": 1341 + }, + { + "epoch": 18.64279475982533, + "grad_norm": 0.01021308358758688, + "learning_rate": 0.002170104590267219, + "loss": 4.485857963562012, + "step": 1342 + }, + { + "epoch": 18.656768558951963, + "grad_norm": 0.012630444951355457, + "learning_rate": 0.00216890810666376, + "loss": 4.442112922668457, + "step": 1343 + }, + { + "epoch": 18.670742358078602, + "grad_norm": 0.013602569699287415, + "learning_rate": 0.0021677110962183515, + "loss": 4.45171594619751, + "step": 1344 + }, + { + "epoch": 18.68471615720524, + "grad_norm": 0.013808641582727432, + "learning_rate": 0.0021665135598954014, + "loss": 4.468818187713623, + "step": 1345 + }, + { + "epoch": 18.69868995633188, + "grad_norm": 0.011510699987411499, + "learning_rate": 0.002165315498659744, + "loss": 4.439584255218506, + "step": 1346 + }, + { + "epoch": 18.712663755458514, + "grad_norm": 0.010502424091100693, + "learning_rate": 0.0021641169134766353, + "loss": 4.478377342224121, + "step": 1347 + }, + { + "epoch": 18.726637554585153, + "grad_norm": 0.009953407570719719, + "learning_rate": 0.0021629178053117542, + "loss": 4.474404335021973, + "step": 1348 + }, + { + "epoch": 18.74061135371179, + "grad_norm": 0.011158930137753487, + "learning_rate": 0.0021617181751312, + "loss": 4.4549560546875, + "step": 1349 + }, + { + "epoch": 18.754585152838427, + "grad_norm": 0.010431940667331219, + "learning_rate": 0.002160518023901493, + "loss": 4.444911003112793, + "step": 1350 + }, + { + "epoch": 18.768558951965066, + "grad_norm": 0.009100335650146008, + "learning_rate": 0.0021593173525895734, + "loss": 4.428071975708008, + "step": 1351 + }, + { + "epoch": 18.782532751091704, + "grad_norm": 0.009362847544252872, + "learning_rate": 0.0021581161621628013, + "loss": 4.462827682495117, + "step": 1352 + }, + { + "epoch": 18.79650655021834, + "grad_norm": 0.009893129579722881, + "learning_rate": 0.0021569144535889535, + "loss": 4.426912307739258, + "step": 1353 + }, + { + "epoch": 18.810480349344978, + "grad_norm": 0.010524468496441841, + "learning_rate": 0.0021557122278362244, + "loss": 4.435606002807617, + "step": 1354 + }, + { + "epoch": 18.824454148471617, + "grad_norm": 0.010621445253491402, + "learning_rate": 0.002154509485873226, + "loss": 4.454901695251465, + "step": 1355 + }, + { + "epoch": 18.83842794759825, + "grad_norm": 0.010836896486580372, + "learning_rate": 0.0021533062286689855, + "loss": 4.4441728591918945, + "step": 1356 + }, + { + "epoch": 18.85240174672489, + "grad_norm": 0.011403030715882778, + "learning_rate": 0.0021521024571929457, + "loss": 4.483884811401367, + "step": 1357 + }, + { + "epoch": 18.86637554585153, + "grad_norm": 0.012522426433861256, + "learning_rate": 0.0021508981724149643, + "loss": 4.464006423950195, + "step": 1358 + }, + { + "epoch": 18.880349344978168, + "grad_norm": 0.014799179509282112, + "learning_rate": 0.0021496933753053105, + "loss": 4.460402011871338, + "step": 1359 + }, + { + "epoch": 18.894323144104803, + "grad_norm": 0.012453878298401833, + "learning_rate": 0.0021484880668346685, + "loss": 4.456899642944336, + "step": 1360 + }, + { + "epoch": 18.90829694323144, + "grad_norm": 0.012270379811525345, + "learning_rate": 0.002147282247974133, + "loss": 4.477325439453125, + "step": 1361 + }, + { + "epoch": 18.92227074235808, + "grad_norm": 0.01094560045748949, + "learning_rate": 0.0021460759196952105, + "loss": 4.422362327575684, + "step": 1362 + }, + { + "epoch": 18.936244541484715, + "grad_norm": 0.009995176456868649, + "learning_rate": 0.0021448690829698183, + "loss": 4.405127048492432, + "step": 1363 + }, + { + "epoch": 18.950218340611354, + "grad_norm": 0.014935536310076714, + "learning_rate": 0.002143661738770283, + "loss": 4.445840358734131, + "step": 1364 + }, + { + "epoch": 18.964192139737992, + "grad_norm": 0.011492030695080757, + "learning_rate": 0.0021424538880693393, + "loss": 4.459555625915527, + "step": 1365 + }, + { + "epoch": 18.978165938864628, + "grad_norm": 0.008793608285486698, + "learning_rate": 0.002141245531840132, + "loss": 4.436631202697754, + "step": 1366 + }, + { + "epoch": 18.992139737991266, + "grad_norm": 0.008894025348126888, + "learning_rate": 0.0021400366710562103, + "loss": 4.473924160003662, + "step": 1367 + }, + { + "epoch": 19.0, + "grad_norm": 0.010324863716959953, + "learning_rate": 0.0021388273066915327, + "loss": 4.475680351257324, + "step": 1368 + }, + { + "epoch": 19.0, + "eval_loss": 4.688324451446533, + "eval_runtime": 54.4515, + "eval_samples_per_second": 44.847, + "eval_steps_per_second": 1.414, + "step": 1368 + }, + { + "epoch": 19.01397379912664, + "grad_norm": 0.009235628880560398, + "learning_rate": 0.0021376174397204615, + "loss": 4.370064735412598, + "step": 1369 + }, + { + "epoch": 19.027947598253274, + "grad_norm": 0.008818200789391994, + "learning_rate": 0.0021364070711177643, + "loss": 4.371678352355957, + "step": 1370 + }, + { + "epoch": 19.041921397379912, + "grad_norm": 0.010333757847547531, + "learning_rate": 0.0021351962018586147, + "loss": 4.3896379470825195, + "step": 1371 + }, + { + "epoch": 19.05589519650655, + "grad_norm": 0.011938809417188168, + "learning_rate": 0.002133984832918587, + "loss": 4.361343860626221, + "step": 1372 + }, + { + "epoch": 19.069868995633186, + "grad_norm": 0.010634614154696465, + "learning_rate": 0.00213277296527366, + "loss": 4.362449645996094, + "step": 1373 + }, + { + "epoch": 19.083842794759825, + "grad_norm": 0.008545687422156334, + "learning_rate": 0.0021315605999002126, + "loss": 4.3685503005981445, + "step": 1374 + }, + { + "epoch": 19.097816593886463, + "grad_norm": 0.009159795008599758, + "learning_rate": 0.002130347737775027, + "loss": 4.379969596862793, + "step": 1375 + }, + { + "epoch": 19.111790393013102, + "grad_norm": 0.010464153252542019, + "learning_rate": 0.0021291343798752835, + "loss": 4.376113414764404, + "step": 1376 + }, + { + "epoch": 19.125764192139737, + "grad_norm": 0.011010527610778809, + "learning_rate": 0.0021279205271785622, + "loss": 4.390711784362793, + "step": 1377 + }, + { + "epoch": 19.139737991266376, + "grad_norm": 0.010350867174565792, + "learning_rate": 0.002126706180662843, + "loss": 4.399683952331543, + "step": 1378 + }, + { + "epoch": 19.153711790393015, + "grad_norm": 0.009708237834274769, + "learning_rate": 0.002125491341306503, + "loss": 4.363435745239258, + "step": 1379 + }, + { + "epoch": 19.16768558951965, + "grad_norm": 0.010087115690112114, + "learning_rate": 0.002124276010088316, + "loss": 4.366623878479004, + "step": 1380 + }, + { + "epoch": 19.18165938864629, + "grad_norm": 0.009712349623441696, + "learning_rate": 0.002123060187987453, + "loss": 4.360814571380615, + "step": 1381 + }, + { + "epoch": 19.195633187772927, + "grad_norm": 0.010174986906349659, + "learning_rate": 0.0021218438759834785, + "loss": 4.347475051879883, + "step": 1382 + }, + { + "epoch": 19.209606986899562, + "grad_norm": 0.010195785202085972, + "learning_rate": 0.0021206270750563543, + "loss": 4.399251461029053, + "step": 1383 + }, + { + "epoch": 19.2235807860262, + "grad_norm": 0.010786959901452065, + "learning_rate": 0.0021194097861864354, + "loss": 4.397933006286621, + "step": 1384 + }, + { + "epoch": 19.23755458515284, + "grad_norm": 0.011476189829409122, + "learning_rate": 0.002118192010354468, + "loss": 4.3748369216918945, + "step": 1385 + }, + { + "epoch": 19.251528384279474, + "grad_norm": 0.011571013368666172, + "learning_rate": 0.0021169737485415927, + "loss": 4.388349533081055, + "step": 1386 + }, + { + "epoch": 19.265502183406113, + "grad_norm": 0.011140935122966766, + "learning_rate": 0.002115755001729341, + "loss": 4.398018836975098, + "step": 1387 + }, + { + "epoch": 19.27947598253275, + "grad_norm": 0.013216734863817692, + "learning_rate": 0.0021145357708996365, + "loss": 4.406947135925293, + "step": 1388 + }, + { + "epoch": 19.29344978165939, + "grad_norm": 0.013385126367211342, + "learning_rate": 0.0021133160570347893, + "loss": 4.41457462310791, + "step": 1389 + }, + { + "epoch": 19.307423580786025, + "grad_norm": 0.011784231290221214, + "learning_rate": 0.002112095861117502, + "loss": 4.403835296630859, + "step": 1390 + }, + { + "epoch": 19.321397379912664, + "grad_norm": 0.009813498705625534, + "learning_rate": 0.0021108751841308645, + "loss": 4.3865647315979, + "step": 1391 + }, + { + "epoch": 19.335371179039303, + "grad_norm": 0.010297936387360096, + "learning_rate": 0.0021096540270583533, + "loss": 4.395583152770996, + "step": 1392 + }, + { + "epoch": 19.349344978165938, + "grad_norm": 0.011965147219598293, + "learning_rate": 0.002108432390883834, + "loss": 4.385443687438965, + "step": 1393 + }, + { + "epoch": 19.363318777292577, + "grad_norm": 0.012212193571031094, + "learning_rate": 0.0021072102765915556, + "loss": 4.422321319580078, + "step": 1394 + }, + { + "epoch": 19.377292576419215, + "grad_norm": 0.009840975515544415, + "learning_rate": 0.0021059876851661544, + "loss": 4.3568902015686035, + "step": 1395 + }, + { + "epoch": 19.39126637554585, + "grad_norm": 0.013039806857705116, + "learning_rate": 0.002104764617592649, + "loss": 4.314867973327637, + "step": 1396 + }, + { + "epoch": 19.40524017467249, + "grad_norm": 0.015091902576386929, + "learning_rate": 0.0021035410748564444, + "loss": 4.392338752746582, + "step": 1397 + }, + { + "epoch": 19.419213973799128, + "grad_norm": 0.015448705293238163, + "learning_rate": 0.0021023170579433257, + "loss": 4.379609107971191, + "step": 1398 + }, + { + "epoch": 19.433187772925763, + "grad_norm": 0.018393104895949364, + "learning_rate": 0.0021010925678394622, + "loss": 4.418291091918945, + "step": 1399 + }, + { + "epoch": 19.4471615720524, + "grad_norm": 0.01866372860968113, + "learning_rate": 0.002099867605531402, + "loss": 4.4006452560424805, + "step": 1400 + }, + { + "epoch": 19.46113537117904, + "grad_norm": 0.01819106750190258, + "learning_rate": 0.002098642172006077, + "loss": 4.433975696563721, + "step": 1401 + }, + { + "epoch": 19.475109170305675, + "grad_norm": 0.019774602726101875, + "learning_rate": 0.0020974162682507952, + "loss": 4.417821884155273, + "step": 1402 + }, + { + "epoch": 19.489082969432314, + "grad_norm": 0.016885802149772644, + "learning_rate": 0.0020961898952532457, + "loss": 4.397073268890381, + "step": 1403 + }, + { + "epoch": 19.503056768558952, + "grad_norm": 0.015750881284475327, + "learning_rate": 0.0020949630540014956, + "loss": 4.4356608390808105, + "step": 1404 + }, + { + "epoch": 19.51703056768559, + "grad_norm": 0.015696514397859573, + "learning_rate": 0.002093735745483988, + "loss": 4.422449588775635, + "step": 1405 + }, + { + "epoch": 19.531004366812226, + "grad_norm": 0.017195021733641624, + "learning_rate": 0.0020925079706895435, + "loss": 4.443209648132324, + "step": 1406 + }, + { + "epoch": 19.544978165938865, + "grad_norm": 0.015864048153162003, + "learning_rate": 0.0020912797306073577, + "loss": 4.4340691566467285, + "step": 1407 + }, + { + "epoch": 19.558951965065503, + "grad_norm": 0.009857800789177418, + "learning_rate": 0.0020900510262270022, + "loss": 4.37935209274292, + "step": 1408 + }, + { + "epoch": 19.57292576419214, + "grad_norm": 0.013739476911723614, + "learning_rate": 0.0020888218585384216, + "loss": 4.412148475646973, + "step": 1409 + }, + { + "epoch": 19.586899563318777, + "grad_norm": 0.011667690239846706, + "learning_rate": 0.0020875922285319344, + "loss": 4.391700744628906, + "step": 1410 + }, + { + "epoch": 19.600873362445416, + "grad_norm": 0.012323460541665554, + "learning_rate": 0.0020863621371982302, + "loss": 4.423880577087402, + "step": 1411 + }, + { + "epoch": 19.61484716157205, + "grad_norm": 0.010353626683354378, + "learning_rate": 0.0020851315855283724, + "loss": 4.427382469177246, + "step": 1412 + }, + { + "epoch": 19.62882096069869, + "grad_norm": 0.0101751908659935, + "learning_rate": 0.002083900574513794, + "loss": 4.415317535400391, + "step": 1413 + }, + { + "epoch": 19.64279475982533, + "grad_norm": 0.01126549206674099, + "learning_rate": 0.002082669105146298, + "loss": 4.402498245239258, + "step": 1414 + }, + { + "epoch": 19.656768558951963, + "grad_norm": 0.011096738278865814, + "learning_rate": 0.0020814371784180575, + "loss": 4.433567047119141, + "step": 1415 + }, + { + "epoch": 19.670742358078602, + "grad_norm": 0.010944445617496967, + "learning_rate": 0.0020802047953216137, + "loss": 4.406990051269531, + "step": 1416 + }, + { + "epoch": 19.68471615720524, + "grad_norm": 0.008965318091213703, + "learning_rate": 0.0020789719568498747, + "loss": 4.3721771240234375, + "step": 1417 + }, + { + "epoch": 19.69868995633188, + "grad_norm": 0.00828113779425621, + "learning_rate": 0.002077738663996117, + "loss": 4.389828205108643, + "step": 1418 + }, + { + "epoch": 19.712663755458514, + "grad_norm": 0.008599454537034035, + "learning_rate": 0.002076504917753982, + "loss": 4.411396026611328, + "step": 1419 + }, + { + "epoch": 19.726637554585153, + "grad_norm": 0.007713889703154564, + "learning_rate": 0.0020752707191174762, + "loss": 4.355343341827393, + "step": 1420 + }, + { + "epoch": 19.74061135371179, + "grad_norm": 0.00912355538457632, + "learning_rate": 0.0020740360690809723, + "loss": 4.423413276672363, + "step": 1421 + }, + { + "epoch": 19.754585152838427, + "grad_norm": 0.008269775658845901, + "learning_rate": 0.0020728009686392045, + "loss": 4.390920639038086, + "step": 1422 + }, + { + "epoch": 19.768558951965066, + "grad_norm": 0.007461601868271828, + "learning_rate": 0.002071565418787272, + "loss": 4.365501403808594, + "step": 1423 + }, + { + "epoch": 19.782532751091704, + "grad_norm": 0.007951333187520504, + "learning_rate": 0.0020703294205206336, + "loss": 4.378884315490723, + "step": 1424 + }, + { + "epoch": 19.79650655021834, + "grad_norm": 0.007816503755748272, + "learning_rate": 0.002069092974835112, + "loss": 4.4241743087768555, + "step": 1425 + }, + { + "epoch": 19.810480349344978, + "grad_norm": 0.00833756010979414, + "learning_rate": 0.0020678560827268886, + "loss": 4.391102313995361, + "step": 1426 + }, + { + "epoch": 19.824454148471617, + "grad_norm": 0.008241684176027775, + "learning_rate": 0.0020666187451925055, + "loss": 4.374970436096191, + "step": 1427 + }, + { + "epoch": 19.83842794759825, + "grad_norm": 0.008245996199548244, + "learning_rate": 0.002065380963228862, + "loss": 4.380222320556641, + "step": 1428 + }, + { + "epoch": 19.85240174672489, + "grad_norm": 0.007378706708550453, + "learning_rate": 0.0020641427378332183, + "loss": 4.391451358795166, + "step": 1429 + }, + { + "epoch": 19.86637554585153, + "grad_norm": 0.007594958413392305, + "learning_rate": 0.002062904070003189, + "loss": 4.372774124145508, + "step": 1430 + }, + { + "epoch": 19.880349344978168, + "grad_norm": 0.0076752156019210815, + "learning_rate": 0.002061664960736747, + "loss": 4.384668827056885, + "step": 1431 + }, + { + "epoch": 19.894323144104803, + "grad_norm": 0.008727089501917362, + "learning_rate": 0.00206042541103222, + "loss": 4.363401412963867, + "step": 1432 + }, + { + "epoch": 19.90829694323144, + "grad_norm": 0.007616396062076092, + "learning_rate": 0.002059185421888291, + "loss": 4.413188934326172, + "step": 1433 + }, + { + "epoch": 19.92227074235808, + "grad_norm": 0.008922954089939594, + "learning_rate": 0.0020579449943039967, + "loss": 4.383417129516602, + "step": 1434 + }, + { + "epoch": 19.936244541484715, + "grad_norm": 0.009063739329576492, + "learning_rate": 0.002056704129278727, + "loss": 4.385283470153809, + "step": 1435 + }, + { + "epoch": 19.950218340611354, + "grad_norm": 0.008912574499845505, + "learning_rate": 0.002055462827812225, + "loss": 4.378478050231934, + "step": 1436 + }, + { + "epoch": 19.964192139737992, + "grad_norm": 0.008317525498569012, + "learning_rate": 0.0020542210909045844, + "loss": 4.367859840393066, + "step": 1437 + }, + { + "epoch": 19.978165938864628, + "grad_norm": 0.009347227402031422, + "learning_rate": 0.0020529789195562506, + "loss": 4.38238000869751, + "step": 1438 + }, + { + "epoch": 19.992139737991266, + "grad_norm": 0.011030065827071667, + "learning_rate": 0.0020517363147680183, + "loss": 4.372624397277832, + "step": 1439 + }, + { + "epoch": 20.0, + "grad_norm": 0.011519700288772583, + "learning_rate": 0.002050493277541032, + "loss": 4.3906121253967285, + "step": 1440 + }, + { + "epoch": 20.0, + "eval_loss": 4.641542434692383, + "eval_runtime": 51.6342, + "eval_samples_per_second": 47.294, + "eval_steps_per_second": 1.491, + "step": 1440 + }, + { + "epoch": 20.01397379912664, + "grad_norm": 0.009694516658782959, + "learning_rate": 0.0020492498088767836, + "loss": 4.32808256149292, + "step": 1441 + }, + { + "epoch": 20.027947598253274, + "grad_norm": 0.0097952950745821, + "learning_rate": 0.002048005909777114, + "loss": 4.280096054077148, + "step": 1442 + }, + { + "epoch": 20.041921397379912, + "grad_norm": 0.010393359698355198, + "learning_rate": 0.0020467615812442106, + "loss": 4.317824363708496, + "step": 1443 + }, + { + "epoch": 20.05589519650655, + "grad_norm": 0.00977372657507658, + "learning_rate": 0.0020455168242806054, + "loss": 4.3445329666137695, + "step": 1444 + }, + { + "epoch": 20.069868995633186, + "grad_norm": 0.008568719029426575, + "learning_rate": 0.0020442716398891775, + "loss": 4.318026542663574, + "step": 1445 + }, + { + "epoch": 20.083842794759825, + "grad_norm": 0.009732510894536972, + "learning_rate": 0.0020430260290731487, + "loss": 4.330609321594238, + "step": 1446 + }, + { + "epoch": 20.097816593886463, + "grad_norm": 0.010429969057440758, + "learning_rate": 0.0020417799928360864, + "loss": 4.316794395446777, + "step": 1447 + }, + { + "epoch": 20.111790393013102, + "grad_norm": 0.009292383678257465, + "learning_rate": 0.002040533532181898, + "loss": 4.317420959472656, + "step": 1448 + }, + { + "epoch": 20.125764192139737, + "grad_norm": 0.011222455650568008, + "learning_rate": 0.0020392866481148356, + "loss": 4.292610168457031, + "step": 1449 + }, + { + "epoch": 20.139737991266376, + "grad_norm": 0.013591725379228592, + "learning_rate": 0.0020380393416394894, + "loss": 4.325689315795898, + "step": 1450 + }, + { + "epoch": 20.153711790393015, + "grad_norm": 0.013572026044130325, + "learning_rate": 0.002036791613760794, + "loss": 4.334960460662842, + "step": 1451 + }, + { + "epoch": 20.16768558951965, + "grad_norm": 0.012986521236598492, + "learning_rate": 0.0020355434654840197, + "loss": 4.33829402923584, + "step": 1452 + }, + { + "epoch": 20.18165938864629, + "grad_norm": 0.013321424834430218, + "learning_rate": 0.0020342948978147777, + "loss": 4.2926926612854, + "step": 1453 + }, + { + "epoch": 20.195633187772927, + "grad_norm": 0.012307215481996536, + "learning_rate": 0.0020330459117590153, + "loss": 4.32757568359375, + "step": 1454 + }, + { + "epoch": 20.209606986899562, + "grad_norm": 0.012909360229969025, + "learning_rate": 0.0020317965083230197, + "loss": 4.362889289855957, + "step": 1455 + }, + { + "epoch": 20.2235807860262, + "grad_norm": 0.017746610566973686, + "learning_rate": 0.0020305466885134115, + "loss": 4.3426313400268555, + "step": 1456 + }, + { + "epoch": 20.23755458515284, + "grad_norm": 0.01696564070880413, + "learning_rate": 0.002029296453337148, + "loss": 4.3298516273498535, + "step": 1457 + }, + { + "epoch": 20.251528384279474, + "grad_norm": 0.013348255306482315, + "learning_rate": 0.002028045803801522, + "loss": 4.360220432281494, + "step": 1458 + }, + { + "epoch": 20.265502183406113, + "grad_norm": 0.013225145637989044, + "learning_rate": 0.002026794740914158, + "loss": 4.346250057220459, + "step": 1459 + }, + { + "epoch": 20.27947598253275, + "grad_norm": 0.01569020375609398, + "learning_rate": 0.002025543265683016, + "loss": 4.308025360107422, + "step": 1460 + }, + { + "epoch": 20.29344978165939, + "grad_norm": 0.018612321466207504, + "learning_rate": 0.002024291379116385, + "loss": 4.330190181732178, + "step": 1461 + }, + { + "epoch": 20.307423580786025, + "grad_norm": 0.01619943603873253, + "learning_rate": 0.00202303908222289, + "loss": 4.339552402496338, + "step": 1462 + }, + { + "epoch": 20.321397379912664, + "grad_norm": 0.015385213308036327, + "learning_rate": 0.0020217863760114826, + "loss": 4.3412184715271, + "step": 1463 + }, + { + "epoch": 20.335371179039303, + "grad_norm": 0.018517112359404564, + "learning_rate": 0.002020533261491446, + "loss": 4.3289313316345215, + "step": 1464 + }, + { + "epoch": 20.349344978165938, + "grad_norm": 0.014868509024381638, + "learning_rate": 0.0020192797396723903, + "loss": 4.359846115112305, + "step": 1465 + }, + { + "epoch": 20.363318777292577, + "grad_norm": 0.014240248128771782, + "learning_rate": 0.0020180258115642578, + "loss": 4.369707107543945, + "step": 1466 + }, + { + "epoch": 20.377292576419215, + "grad_norm": 0.018113158643245697, + "learning_rate": 0.0020167714781773143, + "loss": 4.362668514251709, + "step": 1467 + }, + { + "epoch": 20.39126637554585, + "grad_norm": 0.01802460104227066, + "learning_rate": 0.002015516740522154, + "loss": 4.344742298126221, + "step": 1468 + }, + { + "epoch": 20.40524017467249, + "grad_norm": 0.01414549071341753, + "learning_rate": 0.0020142615996096964, + "loss": 4.358080863952637, + "step": 1469 + }, + { + "epoch": 20.419213973799128, + "grad_norm": 0.012157700955867767, + "learning_rate": 0.002013006056451186, + "loss": 4.347763538360596, + "step": 1470 + }, + { + "epoch": 20.433187772925763, + "grad_norm": 0.011268534697592258, + "learning_rate": 0.0020117501120581905, + "loss": 4.335859298706055, + "step": 1471 + }, + { + "epoch": 20.4471615720524, + "grad_norm": 0.010395628400146961, + "learning_rate": 0.002010493767442602, + "loss": 4.341609001159668, + "step": 1472 + }, + { + "epoch": 20.46113537117904, + "grad_norm": 0.012147950939834118, + "learning_rate": 0.002009237023616635, + "loss": 4.328502655029297, + "step": 1473 + }, + { + "epoch": 20.475109170305675, + "grad_norm": 0.00945813488215208, + "learning_rate": 0.0020079798815928252, + "loss": 4.311186790466309, + "step": 1474 + }, + { + "epoch": 20.489082969432314, + "grad_norm": 0.009489500895142555, + "learning_rate": 0.0020067223423840294, + "loss": 4.371470928192139, + "step": 1475 + }, + { + "epoch": 20.503056768558952, + "grad_norm": 0.009360029362142086, + "learning_rate": 0.0020054644070034234, + "loss": 4.33941125869751, + "step": 1476 + }, + { + "epoch": 20.51703056768559, + "grad_norm": 0.010032779537141323, + "learning_rate": 0.0020042060764645037, + "loss": 4.365797519683838, + "step": 1477 + }, + { + "epoch": 20.531004366812226, + "grad_norm": 0.009367046877741814, + "learning_rate": 0.002002947351781084, + "loss": 4.29443359375, + "step": 1478 + }, + { + "epoch": 20.544978165938865, + "grad_norm": 0.008369733579456806, + "learning_rate": 0.0020016882339672967, + "loss": 4.294233322143555, + "step": 1479 + }, + { + "epoch": 20.558951965065503, + "grad_norm": 0.010394785553216934, + "learning_rate": 0.0020004287240375893, + "loss": 4.352355003356934, + "step": 1480 + }, + { + "epoch": 20.57292576419214, + "grad_norm": 0.01068816613405943, + "learning_rate": 0.001999168823006728, + "loss": 4.342516899108887, + "step": 1481 + }, + { + "epoch": 20.586899563318777, + "grad_norm": 0.009975973516702652, + "learning_rate": 0.0019979085318897905, + "loss": 4.346763610839844, + "step": 1482 + }, + { + "epoch": 20.600873362445416, + "grad_norm": 0.009618151932954788, + "learning_rate": 0.001996647851702171, + "loss": 4.329115867614746, + "step": 1483 + }, + { + "epoch": 20.61484716157205, + "grad_norm": 0.008399694226682186, + "learning_rate": 0.001995386783459577, + "loss": 4.28736686706543, + "step": 1484 + }, + { + "epoch": 20.62882096069869, + "grad_norm": 0.008235268294811249, + "learning_rate": 0.001994125328178029, + "loss": 4.339872360229492, + "step": 1485 + }, + { + "epoch": 20.64279475982533, + "grad_norm": 0.00836547277867794, + "learning_rate": 0.0019928634868738577, + "loss": 4.328156471252441, + "step": 1486 + }, + { + "epoch": 20.656768558951963, + "grad_norm": 0.009619626216590405, + "learning_rate": 0.001991601260563706, + "loss": 4.337394714355469, + "step": 1487 + }, + { + "epoch": 20.670742358078602, + "grad_norm": 0.009814038872718811, + "learning_rate": 0.0019903386502645275, + "loss": 4.325926780700684, + "step": 1488 + }, + { + "epoch": 20.68471615720524, + "grad_norm": 0.009588954970240593, + "learning_rate": 0.001989075656993584, + "loss": 4.286477088928223, + "step": 1489 + }, + { + "epoch": 20.69868995633188, + "grad_norm": 0.010471943765878677, + "learning_rate": 0.001987812281768447, + "loss": 4.322466850280762, + "step": 1490 + }, + { + "epoch": 20.712663755458514, + "grad_norm": 0.010239035822451115, + "learning_rate": 0.001986548525606994, + "loss": 4.3487653732299805, + "step": 1491 + }, + { + "epoch": 20.726637554585153, + "grad_norm": 0.008935568854212761, + "learning_rate": 0.001985284389527411, + "loss": 4.303601264953613, + "step": 1492 + }, + { + "epoch": 20.74061135371179, + "grad_norm": 0.008624759502708912, + "learning_rate": 0.00198401987454819, + "loss": 4.34036922454834, + "step": 1493 + }, + { + "epoch": 20.754585152838427, + "grad_norm": 0.009296547621488571, + "learning_rate": 0.0019827549816881277, + "loss": 4.273357391357422, + "step": 1494 + }, + { + "epoch": 20.768558951965066, + "grad_norm": 0.009071138687431812, + "learning_rate": 0.0019814897119663248, + "loss": 4.321051597595215, + "step": 1495 + }, + { + "epoch": 20.782532751091704, + "grad_norm": 0.008341389708220959, + "learning_rate": 0.0019802240664021876, + "loss": 4.343242645263672, + "step": 1496 + }, + { + "epoch": 20.79650655021834, + "grad_norm": 0.007881715893745422, + "learning_rate": 0.0019789580460154227, + "loss": 4.32067346572876, + "step": 1497 + }, + { + "epoch": 20.810480349344978, + "grad_norm": 0.008750553242862225, + "learning_rate": 0.0019776916518260404, + "loss": 4.330263614654541, + "step": 1498 + }, + { + "epoch": 20.824454148471617, + "grad_norm": 0.008591460064053535, + "learning_rate": 0.001976424884854351, + "loss": 4.311957359313965, + "step": 1499 + }, + { + "epoch": 20.83842794759825, + "grad_norm": 0.008084543980658054, + "learning_rate": 0.001975157746120966, + "loss": 4.331319808959961, + "step": 1500 + }, + { + "epoch": 20.85240174672489, + "grad_norm": 0.008036594837903976, + "learning_rate": 0.001973890236646797, + "loss": 4.382656097412109, + "step": 1501 + }, + { + "epoch": 20.86637554585153, + "grad_norm": 0.00805772002786398, + "learning_rate": 0.0019726223574530527, + "loss": 4.297247886657715, + "step": 1502 + }, + { + "epoch": 20.880349344978168, + "grad_norm": 0.009019420482218266, + "learning_rate": 0.0019713541095612405, + "loss": 4.326526165008545, + "step": 1503 + }, + { + "epoch": 20.894323144104803, + "grad_norm": 0.009953883476555347, + "learning_rate": 0.0019700854939931653, + "loss": 4.327356338500977, + "step": 1504 + }, + { + "epoch": 20.90829694323144, + "grad_norm": 0.009340492077171803, + "learning_rate": 0.0019688165117709273, + "loss": 4.343993186950684, + "step": 1505 + }, + { + "epoch": 20.92227074235808, + "grad_norm": 0.010395721532404423, + "learning_rate": 0.0019675471639169226, + "loss": 4.284934997558594, + "step": 1506 + }, + { + "epoch": 20.936244541484715, + "grad_norm": 0.009946230798959732, + "learning_rate": 0.0019662774514538415, + "loss": 4.288936614990234, + "step": 1507 + }, + { + "epoch": 20.950218340611354, + "grad_norm": 0.010981100611388683, + "learning_rate": 0.0019650073754046695, + "loss": 4.339935779571533, + "step": 1508 + }, + { + "epoch": 20.964192139737992, + "grad_norm": 0.009688783437013626, + "learning_rate": 0.0019637369367926827, + "loss": 4.381425857543945, + "step": 1509 + }, + { + "epoch": 20.978165938864628, + "grad_norm": 0.009442255832254887, + "learning_rate": 0.001962466136641452, + "loss": 4.304185390472412, + "step": 1510 + }, + { + "epoch": 20.992139737991266, + "grad_norm": 0.009371291846036911, + "learning_rate": 0.0019611949759748364, + "loss": 4.306120872497559, + "step": 1511 + }, + { + "epoch": 21.0, + "grad_norm": 0.009544814005494118, + "learning_rate": 0.001959923455816988, + "loss": 4.326251983642578, + "step": 1512 + }, + { + "epoch": 21.0, + "eval_loss": 4.602745532989502, + "eval_runtime": 53.5041, + "eval_samples_per_second": 45.641, + "eval_steps_per_second": 1.439, + "step": 1512 + }, + { + "epoch": 21.01397379912664, + "grad_norm": 0.01010364294052124, + "learning_rate": 0.0019586515771923474, + "loss": 4.24491024017334, + "step": 1513 + }, + { + "epoch": 21.027947598253274, + "grad_norm": 0.012722729705274105, + "learning_rate": 0.0019573793411256446, + "loss": 4.257739543914795, + "step": 1514 + }, + { + "epoch": 21.041921397379912, + "grad_norm": 0.01307204831391573, + "learning_rate": 0.0019561067486418967, + "loss": 4.291684150695801, + "step": 1515 + }, + { + "epoch": 21.05589519650655, + "grad_norm": 0.011215966194868088, + "learning_rate": 0.0019548338007664098, + "loss": 4.237947463989258, + "step": 1516 + }, + { + "epoch": 21.069868995633186, + "grad_norm": 0.012943019159138203, + "learning_rate": 0.0019535604985247727, + "loss": 4.242859363555908, + "step": 1517 + }, + { + "epoch": 21.083842794759825, + "grad_norm": 0.01337282732129097, + "learning_rate": 0.0019522868429428636, + "loss": 4.238934516906738, + "step": 1518 + }, + { + "epoch": 21.097816593886463, + "grad_norm": 0.013697991147637367, + "learning_rate": 0.0019510128350468438, + "loss": 4.285760402679443, + "step": 1519 + }, + { + "epoch": 21.111790393013102, + "grad_norm": 0.012182159349322319, + "learning_rate": 0.001949738475863158, + "loss": 4.290044784545898, + "step": 1520 + }, + { + "epoch": 21.125764192139737, + "grad_norm": 0.012620339170098305, + "learning_rate": 0.0019484637664185346, + "loss": 4.267187118530273, + "step": 1521 + }, + { + "epoch": 21.139737991266376, + "grad_norm": 0.011284958571195602, + "learning_rate": 0.0019471887077399837, + "loss": 4.26662540435791, + "step": 1522 + }, + { + "epoch": 21.153711790393015, + "grad_norm": 0.01161450706422329, + "learning_rate": 0.001945913300854797, + "loss": 4.283192157745361, + "step": 1523 + }, + { + "epoch": 21.16768558951965, + "grad_norm": 0.010022968985140324, + "learning_rate": 0.001944637546790548, + "loss": 4.295843124389648, + "step": 1524 + }, + { + "epoch": 21.18165938864629, + "grad_norm": 0.010341878980398178, + "learning_rate": 0.001943361446575087, + "loss": 4.295636177062988, + "step": 1525 + }, + { + "epoch": 21.195633187772927, + "grad_norm": 0.011549185030162334, + "learning_rate": 0.0019420850012365457, + "loss": 4.288768768310547, + "step": 1526 + }, + { + "epoch": 21.209606986899562, + "grad_norm": 0.012518810108304024, + "learning_rate": 0.0019408082118033331, + "loss": 4.229338645935059, + "step": 1527 + }, + { + "epoch": 21.2235807860262, + "grad_norm": 0.012035328894853592, + "learning_rate": 0.0019395310793041353, + "loss": 4.258639335632324, + "step": 1528 + }, + { + "epoch": 21.23755458515284, + "grad_norm": 0.011193782091140747, + "learning_rate": 0.0019382536047679146, + "loss": 4.239264488220215, + "step": 1529 + }, + { + "epoch": 21.251528384279474, + "grad_norm": 0.009930136613547802, + "learning_rate": 0.0019369757892239096, + "loss": 4.252286434173584, + "step": 1530 + }, + { + "epoch": 21.265502183406113, + "grad_norm": 0.00953229982405901, + "learning_rate": 0.0019356976337016327, + "loss": 4.293435096740723, + "step": 1531 + }, + { + "epoch": 21.27947598253275, + "grad_norm": 0.010618340224027634, + "learning_rate": 0.0019344191392308716, + "loss": 4.235588550567627, + "step": 1532 + }, + { + "epoch": 21.29344978165939, + "grad_norm": 0.008806651458144188, + "learning_rate": 0.0019331403068416856, + "loss": 4.26716423034668, + "step": 1533 + }, + { + "epoch": 21.307423580786025, + "grad_norm": 0.008934796787798405, + "learning_rate": 0.0019318611375644065, + "loss": 4.258462905883789, + "step": 1534 + }, + { + "epoch": 21.321397379912664, + "grad_norm": 0.009283283725380898, + "learning_rate": 0.0019305816324296386, + "loss": 4.257868766784668, + "step": 1535 + }, + { + "epoch": 21.335371179039303, + "grad_norm": 0.009224812500178814, + "learning_rate": 0.0019293017924682555, + "loss": 4.258486747741699, + "step": 1536 + }, + { + "epoch": 21.349344978165938, + "grad_norm": 0.008957969956099987, + "learning_rate": 0.0019280216187114016, + "loss": 4.309563636779785, + "step": 1537 + }, + { + "epoch": 21.363318777292577, + "grad_norm": 0.010284051299095154, + "learning_rate": 0.0019267411121904894, + "loss": 4.275271415710449, + "step": 1538 + }, + { + "epoch": 21.377292576419215, + "grad_norm": 0.009167762473225594, + "learning_rate": 0.0019254602739372001, + "loss": 4.238452911376953, + "step": 1539 + }, + { + "epoch": 21.39126637554585, + "grad_norm": 0.009411636739969254, + "learning_rate": 0.0019241791049834816, + "loss": 4.2735595703125, + "step": 1540 + }, + { + "epoch": 21.40524017467249, + "grad_norm": 0.009044171310961246, + "learning_rate": 0.0019228976063615497, + "loss": 4.27017879486084, + "step": 1541 + }, + { + "epoch": 21.419213973799128, + "grad_norm": 0.008840068243443966, + "learning_rate": 0.0019216157791038834, + "loss": 4.28773307800293, + "step": 1542 + }, + { + "epoch": 21.433187772925763, + "grad_norm": 0.009512359276413918, + "learning_rate": 0.001920333624243228, + "loss": 4.266008377075195, + "step": 1543 + }, + { + "epoch": 21.4471615720524, + "grad_norm": 0.009946667589247227, + "learning_rate": 0.001919051142812593, + "loss": 4.279062271118164, + "step": 1544 + }, + { + "epoch": 21.46113537117904, + "grad_norm": 0.00899894442409277, + "learning_rate": 0.0019177683358452505, + "loss": 4.277679443359375, + "step": 1545 + }, + { + "epoch": 21.475109170305675, + "grad_norm": 0.00909668579697609, + "learning_rate": 0.0019164852043747338, + "loss": 4.311397552490234, + "step": 1546 + }, + { + "epoch": 21.489082969432314, + "grad_norm": 0.009787327609956264, + "learning_rate": 0.0019152017494348408, + "loss": 4.30875825881958, + "step": 1547 + }, + { + "epoch": 21.503056768558952, + "grad_norm": 0.008983899839222431, + "learning_rate": 0.001913917972059626, + "loss": 4.259649276733398, + "step": 1548 + }, + { + "epoch": 21.51703056768559, + "grad_norm": 0.008411582559347153, + "learning_rate": 0.0019126338732834071, + "loss": 4.259811878204346, + "step": 1549 + }, + { + "epoch": 21.531004366812226, + "grad_norm": 0.009162918664515018, + "learning_rate": 0.0019113494541407579, + "loss": 4.271515846252441, + "step": 1550 + }, + { + "epoch": 21.544978165938865, + "grad_norm": 0.011304561980068684, + "learning_rate": 0.0019100647156665125, + "loss": 4.269217491149902, + "step": 1551 + }, + { + "epoch": 21.558951965065503, + "grad_norm": 0.012483394704759121, + "learning_rate": 0.0019087796588957617, + "loss": 4.2949299812316895, + "step": 1552 + }, + { + "epoch": 21.57292576419214, + "grad_norm": 0.012091380544006824, + "learning_rate": 0.0019074942848638522, + "loss": 4.289752006530762, + "step": 1553 + }, + { + "epoch": 21.586899563318777, + "grad_norm": 0.010011567734181881, + "learning_rate": 0.0019062085946063869, + "loss": 4.294095993041992, + "step": 1554 + }, + { + "epoch": 21.600873362445416, + "grad_norm": 0.008766217157244682, + "learning_rate": 0.0019049225891592226, + "loss": 4.256425857543945, + "step": 1555 + }, + { + "epoch": 21.61484716157205, + "grad_norm": 0.012862488627433777, + "learning_rate": 0.001903636269558471, + "loss": 4.287847995758057, + "step": 1556 + }, + { + "epoch": 21.62882096069869, + "grad_norm": 0.013917063362896442, + "learning_rate": 0.0019023496368404972, + "loss": 4.282784938812256, + "step": 1557 + }, + { + "epoch": 21.64279475982533, + "grad_norm": 0.012129023671150208, + "learning_rate": 0.0019010626920419173, + "loss": 4.2486419677734375, + "step": 1558 + }, + { + "epoch": 21.656768558951963, + "grad_norm": 0.010348426178097725, + "learning_rate": 0.0018997754361995986, + "loss": 4.289836883544922, + "step": 1559 + }, + { + "epoch": 21.670742358078602, + "grad_norm": 0.011821693740785122, + "learning_rate": 0.001898487870350662, + "loss": 4.214344024658203, + "step": 1560 + }, + { + "epoch": 21.68471615720524, + "grad_norm": 0.013136724010109901, + "learning_rate": 0.0018971999955324742, + "loss": 4.24208402633667, + "step": 1561 + }, + { + "epoch": 21.69868995633188, + "grad_norm": 0.013173168525099754, + "learning_rate": 0.0018959118127826538, + "loss": 4.270164489746094, + "step": 1562 + }, + { + "epoch": 21.712663755458514, + "grad_norm": 0.011018159799277782, + "learning_rate": 0.0018946233231390656, + "loss": 4.2571916580200195, + "step": 1563 + }, + { + "epoch": 21.726637554585153, + "grad_norm": 0.01018658559769392, + "learning_rate": 0.0018933345276398227, + "loss": 4.283267498016357, + "step": 1564 + }, + { + "epoch": 21.74061135371179, + "grad_norm": 0.009412688203155994, + "learning_rate": 0.0018920454273232848, + "loss": 4.272198677062988, + "step": 1565 + }, + { + "epoch": 21.754585152838427, + "grad_norm": 0.007799267303198576, + "learning_rate": 0.0018907560232280557, + "loss": 4.260125637054443, + "step": 1566 + }, + { + "epoch": 21.768558951965066, + "grad_norm": 0.008747464045882225, + "learning_rate": 0.0018894663163929863, + "loss": 4.2539215087890625, + "step": 1567 + }, + { + "epoch": 21.782532751091704, + "grad_norm": 0.008003434166312218, + "learning_rate": 0.0018881763078571689, + "loss": 4.245024681091309, + "step": 1568 + }, + { + "epoch": 21.79650655021834, + "grad_norm": 0.007593846879899502, + "learning_rate": 0.0018868859986599415, + "loss": 4.237929344177246, + "step": 1569 + }, + { + "epoch": 21.810480349344978, + "grad_norm": 0.008099420927464962, + "learning_rate": 0.0018855953898408815, + "loss": 4.2789106369018555, + "step": 1570 + }, + { + "epoch": 21.824454148471617, + "grad_norm": 0.008270486257970333, + "learning_rate": 0.0018843044824398097, + "loss": 4.263522148132324, + "step": 1571 + }, + { + "epoch": 21.83842794759825, + "grad_norm": 0.009260379709303379, + "learning_rate": 0.0018830132774967866, + "loss": 4.275835037231445, + "step": 1572 + }, + { + "epoch": 21.85240174672489, + "grad_norm": 0.009289492852985859, + "learning_rate": 0.0018817217760521132, + "loss": 4.313778400421143, + "step": 1573 + }, + { + "epoch": 21.86637554585153, + "grad_norm": 0.009493921883404255, + "learning_rate": 0.0018804299791463281, + "loss": 4.303672790527344, + "step": 1574 + }, + { + "epoch": 21.880349344978168, + "grad_norm": 0.009020023047924042, + "learning_rate": 0.0018791378878202105, + "loss": 4.245473384857178, + "step": 1575 + }, + { + "epoch": 21.894323144104803, + "grad_norm": 0.009446556679904461, + "learning_rate": 0.0018778455031147724, + "loss": 4.267327785491943, + "step": 1576 + }, + { + "epoch": 21.90829694323144, + "grad_norm": 0.009499974548816681, + "learning_rate": 0.0018765528260712672, + "loss": 4.301205635070801, + "step": 1577 + }, + { + "epoch": 21.92227074235808, + "grad_norm": 0.008399393409490585, + "learning_rate": 0.00187525985773118, + "loss": 4.307802677154541, + "step": 1578 + }, + { + "epoch": 21.936244541484715, + "grad_norm": 0.009519701823592186, + "learning_rate": 0.0018739665991362327, + "loss": 4.284796714782715, + "step": 1579 + }, + { + "epoch": 21.950218340611354, + "grad_norm": 0.01125232595950365, + "learning_rate": 0.0018726730513283804, + "loss": 4.285930633544922, + "step": 1580 + }, + { + "epoch": 21.964192139737992, + "grad_norm": 0.011608537286520004, + "learning_rate": 0.00187137921534981, + "loss": 4.257689476013184, + "step": 1581 + }, + { + "epoch": 21.978165938864628, + "grad_norm": 0.01064533181488514, + "learning_rate": 0.001870085092242944, + "loss": 4.249063968658447, + "step": 1582 + }, + { + "epoch": 21.992139737991266, + "grad_norm": 0.009854599833488464, + "learning_rate": 0.0018687906830504321, + "loss": 4.255159378051758, + "step": 1583 + }, + { + "epoch": 22.0, + "grad_norm": 0.01121739111840725, + "learning_rate": 0.0018674959888151576, + "loss": 4.287777900695801, + "step": 1584 + }, + { + "epoch": 22.0, + "eval_loss": 4.5675368309021, + "eval_runtime": 51.8555, + "eval_samples_per_second": 47.092, + "eval_steps_per_second": 1.485, + "step": 1584 + }, + { + "epoch": 22.01397379912664, + "grad_norm": 0.0104464590549469, + "learning_rate": 0.0018662010105802314, + "loss": 4.206080436706543, + "step": 1585 + }, + { + "epoch": 22.027947598253274, + "grad_norm": 0.012800181284546852, + "learning_rate": 0.0018649057493889945, + "loss": 4.213550567626953, + "step": 1586 + }, + { + "epoch": 22.041921397379912, + "grad_norm": 0.013382901437580585, + "learning_rate": 0.001863610206285015, + "loss": 4.190858840942383, + "step": 1587 + }, + { + "epoch": 22.05589519650655, + "grad_norm": 0.014347050338983536, + "learning_rate": 0.0018623143823120894, + "loss": 4.22884464263916, + "step": 1588 + }, + { + "epoch": 22.069868995633186, + "grad_norm": 0.01347152329981327, + "learning_rate": 0.001861018278514239, + "loss": 4.215579032897949, + "step": 1589 + }, + { + "epoch": 22.083842794759825, + "grad_norm": 0.012786377221345901, + "learning_rate": 0.0018597218959357112, + "loss": 4.174999237060547, + "step": 1590 + }, + { + "epoch": 22.097816593886463, + "grad_norm": 0.014137202873826027, + "learning_rate": 0.001858425235620979, + "loss": 4.21709680557251, + "step": 1591 + }, + { + "epoch": 22.111790393013102, + "grad_norm": 0.013864399865269661, + "learning_rate": 0.001857128298614737, + "loss": 4.2375030517578125, + "step": 1592 + }, + { + "epoch": 22.125764192139737, + "grad_norm": 0.010397658683359623, + "learning_rate": 0.0018558310859619048, + "loss": 4.231938362121582, + "step": 1593 + }, + { + "epoch": 22.139737991266376, + "grad_norm": 0.00888819433748722, + "learning_rate": 0.0018545335987076223, + "loss": 4.195562839508057, + "step": 1594 + }, + { + "epoch": 22.153711790393015, + "grad_norm": 0.010698067955672741, + "learning_rate": 0.001853235837897252, + "loss": 4.22847318649292, + "step": 1595 + }, + { + "epoch": 22.16768558951965, + "grad_norm": 0.00987589918076992, + "learning_rate": 0.0018519378045763766, + "loss": 4.2147040367126465, + "step": 1596 + }, + { + "epoch": 22.18165938864629, + "grad_norm": 0.011451782658696175, + "learning_rate": 0.001850639499790798, + "loss": 4.179263114929199, + "step": 1597 + }, + { + "epoch": 22.195633187772927, + "grad_norm": 0.011224704794585705, + "learning_rate": 0.0018493409245865368, + "loss": 4.23030948638916, + "step": 1598 + }, + { + "epoch": 22.209606986899562, + "grad_norm": 0.00996861606836319, + "learning_rate": 0.0018480420800098313, + "loss": 4.194283962249756, + "step": 1599 + }, + { + "epoch": 22.2235807860262, + "grad_norm": 0.012143119238317013, + "learning_rate": 0.0018467429671071374, + "loss": 4.221704959869385, + "step": 1600 + }, + { + "epoch": 22.23755458515284, + "grad_norm": 0.013126783072948456, + "learning_rate": 0.0018454435869251268, + "loss": 4.187748908996582, + "step": 1601 + }, + { + "epoch": 22.251528384279474, + "grad_norm": 0.013031953014433384, + "learning_rate": 0.001844143940510687, + "loss": 4.248250961303711, + "step": 1602 + }, + { + "epoch": 22.265502183406113, + "grad_norm": 0.012868993915617466, + "learning_rate": 0.001842844028910919, + "loss": 4.20161247253418, + "step": 1603 + }, + { + "epoch": 22.27947598253275, + "grad_norm": 0.010102041065692902, + "learning_rate": 0.0018415438531731395, + "loss": 4.213654518127441, + "step": 1604 + }, + { + "epoch": 22.29344978165939, + "grad_norm": 0.009599820710718632, + "learning_rate": 0.0018402434143448747, + "loss": 4.206313133239746, + "step": 1605 + }, + { + "epoch": 22.307423580786025, + "grad_norm": 0.011536866426467896, + "learning_rate": 0.0018389427134738658, + "loss": 4.237072944641113, + "step": 1606 + }, + { + "epoch": 22.321397379912664, + "grad_norm": 0.009558089077472687, + "learning_rate": 0.0018376417516080636, + "loss": 4.182195663452148, + "step": 1607 + }, + { + "epoch": 22.335371179039303, + "grad_norm": 0.01056219544261694, + "learning_rate": 0.0018363405297956302, + "loss": 4.232585906982422, + "step": 1608 + }, + { + "epoch": 22.349344978165938, + "grad_norm": 0.012673584744334221, + "learning_rate": 0.0018350390490849354, + "loss": 4.228830337524414, + "step": 1609 + }, + { + "epoch": 22.363318777292577, + "grad_norm": 0.011215884238481522, + "learning_rate": 0.0018337373105245598, + "loss": 4.216056823730469, + "step": 1610 + }, + { + "epoch": 22.377292576419215, + "grad_norm": 0.011281915940344334, + "learning_rate": 0.00183243531516329, + "loss": 4.233404159545898, + "step": 1611 + }, + { + "epoch": 22.39126637554585, + "grad_norm": 0.010381227359175682, + "learning_rate": 0.0018311330640501204, + "loss": 4.182013034820557, + "step": 1612 + }, + { + "epoch": 22.40524017467249, + "grad_norm": 0.00900198519229889, + "learning_rate": 0.0018298305582342513, + "loss": 4.215488433837891, + "step": 1613 + }, + { + "epoch": 22.419213973799128, + "grad_norm": 0.010198601521551609, + "learning_rate": 0.0018285277987650876, + "loss": 4.19389533996582, + "step": 1614 + }, + { + "epoch": 22.433187772925763, + "grad_norm": 0.009965402074158192, + "learning_rate": 0.0018272247866922401, + "loss": 4.218005657196045, + "step": 1615 + }, + { + "epoch": 22.4471615720524, + "grad_norm": 0.009299935773015022, + "learning_rate": 0.0018259215230655202, + "loss": 4.223684310913086, + "step": 1616 + }, + { + "epoch": 22.46113537117904, + "grad_norm": 0.008977951481938362, + "learning_rate": 0.0018246180089349462, + "loss": 4.215778350830078, + "step": 1617 + }, + { + "epoch": 22.475109170305675, + "grad_norm": 0.009261307306587696, + "learning_rate": 0.0018233142453507344, + "loss": 4.215066909790039, + "step": 1618 + }, + { + "epoch": 22.489082969432314, + "grad_norm": 0.008322088047862053, + "learning_rate": 0.0018220102333633044, + "loss": 4.2188005447387695, + "step": 1619 + }, + { + "epoch": 22.503056768558952, + "grad_norm": 0.008716144599020481, + "learning_rate": 0.0018207059740232739, + "loss": 4.187867641448975, + "step": 1620 + }, + { + "epoch": 22.51703056768559, + "grad_norm": 0.008897020481526852, + "learning_rate": 0.0018194014683814625, + "loss": 4.218034267425537, + "step": 1621 + }, + { + "epoch": 22.531004366812226, + "grad_norm": 0.007298174314200878, + "learning_rate": 0.0018180967174888857, + "loss": 4.2017011642456055, + "step": 1622 + }, + { + "epoch": 22.544978165938865, + "grad_norm": 0.008491789922118187, + "learning_rate": 0.0018167917223967582, + "loss": 4.239873886108398, + "step": 1623 + }, + { + "epoch": 22.558951965065503, + "grad_norm": 0.008413378149271011, + "learning_rate": 0.0018154864841564908, + "loss": 4.234771251678467, + "step": 1624 + }, + { + "epoch": 22.57292576419214, + "grad_norm": 0.008141471073031425, + "learning_rate": 0.0018141810038196909, + "loss": 4.185944080352783, + "step": 1625 + }, + { + "epoch": 22.586899563318777, + "grad_norm": 0.009480944834649563, + "learning_rate": 0.0018128752824381603, + "loss": 4.206586837768555, + "step": 1626 + }, + { + "epoch": 22.600873362445416, + "grad_norm": 0.0106047997251153, + "learning_rate": 0.001811569321063894, + "loss": 4.235287666320801, + "step": 1627 + }, + { + "epoch": 22.61484716157205, + "grad_norm": 0.010663514956831932, + "learning_rate": 0.0018102631207490838, + "loss": 4.204899311065674, + "step": 1628 + }, + { + "epoch": 22.62882096069869, + "grad_norm": 0.011473523452877998, + "learning_rate": 0.001808956682546109, + "loss": 4.253668308258057, + "step": 1629 + }, + { + "epoch": 22.64279475982533, + "grad_norm": 0.01165992021560669, + "learning_rate": 0.0018076500075075456, + "loss": 4.209845542907715, + "step": 1630 + }, + { + "epoch": 22.656768558951963, + "grad_norm": 0.011692410334944725, + "learning_rate": 0.0018063430966861564, + "loss": 4.250368595123291, + "step": 1631 + }, + { + "epoch": 22.670742358078602, + "grad_norm": 0.01061506848782301, + "learning_rate": 0.0018050359511348965, + "loss": 4.193821907043457, + "step": 1632 + }, + { + "epoch": 22.68471615720524, + "grad_norm": 0.010510941036045551, + "learning_rate": 0.0018037285719069097, + "loss": 4.254327774047852, + "step": 1633 + }, + { + "epoch": 22.69868995633188, + "grad_norm": 0.011587041430175304, + "learning_rate": 0.001802420960055528, + "loss": 4.2838945388793945, + "step": 1634 + }, + { + "epoch": 22.712663755458514, + "grad_norm": 0.01257420051842928, + "learning_rate": 0.0018011131166342693, + "loss": 4.258241653442383, + "step": 1635 + }, + { + "epoch": 22.726637554585153, + "grad_norm": 0.00936233438551426, + "learning_rate": 0.0017998050426968407, + "loss": 4.201854705810547, + "step": 1636 + }, + { + "epoch": 22.74061135371179, + "grad_norm": 0.010093127377331257, + "learning_rate": 0.0017984967392971338, + "loss": 4.193051338195801, + "step": 1637 + }, + { + "epoch": 22.754585152838427, + "grad_norm": 0.010032009333372116, + "learning_rate": 0.0017971882074892236, + "loss": 4.230212211608887, + "step": 1638 + }, + { + "epoch": 22.768558951965066, + "grad_norm": 0.009455587714910507, + "learning_rate": 0.0017958794483273726, + "loss": 4.2157673835754395, + "step": 1639 + }, + { + "epoch": 22.782532751091704, + "grad_norm": 0.009549238719046116, + "learning_rate": 0.001794570462866023, + "loss": 4.260272026062012, + "step": 1640 + }, + { + "epoch": 22.79650655021834, + "grad_norm": 0.011059699580073357, + "learning_rate": 0.0017932612521598013, + "loss": 4.214818000793457, + "step": 1641 + }, + { + "epoch": 22.810480349344978, + "grad_norm": 0.011653627268970013, + "learning_rate": 0.0017919518172635146, + "loss": 4.205112457275391, + "step": 1642 + }, + { + "epoch": 22.824454148471617, + "grad_norm": 0.012438314035534859, + "learning_rate": 0.001790642159232151, + "loss": 4.179045677185059, + "step": 1643 + }, + { + "epoch": 22.83842794759825, + "grad_norm": 0.011234673671424389, + "learning_rate": 0.0017893322791208784, + "loss": 4.239317417144775, + "step": 1644 + }, + { + "epoch": 22.85240174672489, + "grad_norm": 0.010923845693469048, + "learning_rate": 0.0017880221779850434, + "loss": 4.230571269989014, + "step": 1645 + }, + { + "epoch": 22.86637554585153, + "grad_norm": 0.009219309315085411, + "learning_rate": 0.001786711856880171, + "loss": 4.229068756103516, + "step": 1646 + }, + { + "epoch": 22.880349344978168, + "grad_norm": 0.009849398396909237, + "learning_rate": 0.0017854013168619638, + "loss": 4.254772186279297, + "step": 1647 + }, + { + "epoch": 22.894323144104803, + "grad_norm": 0.01117763202637434, + "learning_rate": 0.001784090558986299, + "loss": 4.252647399902344, + "step": 1648 + }, + { + "epoch": 22.90829694323144, + "grad_norm": 0.010664612054824829, + "learning_rate": 0.0017827795843092316, + "loss": 4.204596519470215, + "step": 1649 + }, + { + "epoch": 22.92227074235808, + "grad_norm": 0.00847950391471386, + "learning_rate": 0.0017814683938869895, + "loss": 4.211061954498291, + "step": 1650 + }, + { + "epoch": 22.936244541484715, + "grad_norm": 0.007305726408958435, + "learning_rate": 0.0017801569887759753, + "loss": 4.181221008300781, + "step": 1651 + }, + { + "epoch": 22.950218340611354, + "grad_norm": 0.007970666512846947, + "learning_rate": 0.0017788453700327653, + "loss": 4.237855434417725, + "step": 1652 + }, + { + "epoch": 22.964192139737992, + "grad_norm": 0.008873614482581615, + "learning_rate": 0.0017775335387141054, + "loss": 4.229188919067383, + "step": 1653 + }, + { + "epoch": 22.978165938864628, + "grad_norm": 0.009691286832094193, + "learning_rate": 0.0017762214958769156, + "loss": 4.205491542816162, + "step": 1654 + }, + { + "epoch": 22.992139737991266, + "grad_norm": 0.009360672906041145, + "learning_rate": 0.0017749092425782852, + "loss": 4.233466625213623, + "step": 1655 + }, + { + "epoch": 23.0, + "grad_norm": 0.009895884431898594, + "learning_rate": 0.0017735967798754723, + "loss": 4.187763214111328, + "step": 1656 + }, + { + "epoch": 23.0, + "eval_loss": 4.535672187805176, + "eval_runtime": 53.1736, + "eval_samples_per_second": 45.925, + "eval_steps_per_second": 1.448, + "step": 1656 + }, + { + "epoch": 23.01397379912664, + "grad_norm": 0.00966129545122385, + "learning_rate": 0.0017722841088259045, + "loss": 4.1451735496521, + "step": 1657 + }, + { + "epoch": 23.027947598253274, + "grad_norm": 0.010733380913734436, + "learning_rate": 0.0017709712304871774, + "loss": 4.1616363525390625, + "step": 1658 + }, + { + "epoch": 23.041921397379912, + "grad_norm": 0.0117328567430377, + "learning_rate": 0.0017696581459170535, + "loss": 4.126331806182861, + "step": 1659 + }, + { + "epoch": 23.05589519650655, + "grad_norm": 0.012640298344194889, + "learning_rate": 0.0017683448561734607, + "loss": 4.163730621337891, + "step": 1660 + }, + { + "epoch": 23.069868995633186, + "grad_norm": 0.013074076734483242, + "learning_rate": 0.0017670313623144934, + "loss": 4.199502944946289, + "step": 1661 + }, + { + "epoch": 23.083842794759825, + "grad_norm": 0.01132719125598669, + "learning_rate": 0.0017657176653984097, + "loss": 4.143380165100098, + "step": 1662 + }, + { + "epoch": 23.097816593886463, + "grad_norm": 0.010008333250880241, + "learning_rate": 0.0017644037664836319, + "loss": 4.170320510864258, + "step": 1663 + }, + { + "epoch": 23.111790393013102, + "grad_norm": 0.011090578511357307, + "learning_rate": 0.0017630896666287438, + "loss": 4.1571478843688965, + "step": 1664 + }, + { + "epoch": 23.125764192139737, + "grad_norm": 0.011047456413507462, + "learning_rate": 0.0017617753668924924, + "loss": 4.101892948150635, + "step": 1665 + }, + { + "epoch": 23.139737991266376, + "grad_norm": 0.010098653845489025, + "learning_rate": 0.0017604608683337857, + "loss": 4.154077529907227, + "step": 1666 + }, + { + "epoch": 23.153711790393015, + "grad_norm": 0.010499860160052776, + "learning_rate": 0.00175914617201169, + "loss": 4.116868019104004, + "step": 1667 + }, + { + "epoch": 23.16768558951965, + "grad_norm": 0.010939318686723709, + "learning_rate": 0.0017578312789854338, + "loss": 4.159144878387451, + "step": 1668 + }, + { + "epoch": 23.18165938864629, + "grad_norm": 0.010622999630868435, + "learning_rate": 0.0017565161903144023, + "loss": 4.171411991119385, + "step": 1669 + }, + { + "epoch": 23.195633187772927, + "grad_norm": 0.0108700692653656, + "learning_rate": 0.0017552009070581386, + "loss": 4.100748062133789, + "step": 1670 + }, + { + "epoch": 23.209606986899562, + "grad_norm": 0.009794988669455051, + "learning_rate": 0.0017538854302763423, + "loss": 4.185161590576172, + "step": 1671 + }, + { + "epoch": 23.2235807860262, + "grad_norm": 0.007750354241579771, + "learning_rate": 0.0017525697610288702, + "loss": 4.149083137512207, + "step": 1672 + }, + { + "epoch": 23.23755458515284, + "grad_norm": 0.009923134930431843, + "learning_rate": 0.001751253900375732, + "loss": 4.205166816711426, + "step": 1673 + }, + { + "epoch": 23.251528384279474, + "grad_norm": 0.010811141692101955, + "learning_rate": 0.0017499378493770936, + "loss": 4.131865978240967, + "step": 1674 + }, + { + "epoch": 23.265502183406113, + "grad_norm": 0.010868688113987446, + "learning_rate": 0.0017486216090932733, + "loss": 4.148791790008545, + "step": 1675 + }, + { + "epoch": 23.27947598253275, + "grad_norm": 0.011224126443266869, + "learning_rate": 0.001747305180584743, + "loss": 4.152503490447998, + "step": 1676 + }, + { + "epoch": 23.29344978165939, + "grad_norm": 0.013262554071843624, + "learning_rate": 0.001745988564912124, + "loss": 4.164926528930664, + "step": 1677 + }, + { + "epoch": 23.307423580786025, + "grad_norm": 0.013089599087834358, + "learning_rate": 0.001744671763136191, + "loss": 4.1626458168029785, + "step": 1678 + }, + { + "epoch": 23.321397379912664, + "grad_norm": 0.01327646616846323, + "learning_rate": 0.0017433547763178665, + "loss": 4.189356803894043, + "step": 1679 + }, + { + "epoch": 23.335371179039303, + "grad_norm": 0.014539042487740517, + "learning_rate": 0.0017420376055182233, + "loss": 4.176387786865234, + "step": 1680 + }, + { + "epoch": 23.349344978165938, + "grad_norm": 0.013269098475575447, + "learning_rate": 0.001740720251798482, + "loss": 4.180716514587402, + "step": 1681 + }, + { + "epoch": 23.363318777292577, + "grad_norm": 0.010922391898930073, + "learning_rate": 0.0017394027162200105, + "loss": 4.163928031921387, + "step": 1682 + }, + { + "epoch": 23.377292576419215, + "grad_norm": 0.013355271890759468, + "learning_rate": 0.0017380849998443243, + "loss": 4.150012016296387, + "step": 1683 + }, + { + "epoch": 23.39126637554585, + "grad_norm": 0.014149201102554798, + "learning_rate": 0.0017367671037330828, + "loss": 4.173713207244873, + "step": 1684 + }, + { + "epoch": 23.40524017467249, + "grad_norm": 0.01354879979044199, + "learning_rate": 0.0017354490289480918, + "loss": 4.150563716888428, + "step": 1685 + }, + { + "epoch": 23.419213973799128, + "grad_norm": 0.013788704760372639, + "learning_rate": 0.0017341307765512998, + "loss": 4.213858604431152, + "step": 1686 + }, + { + "epoch": 23.433187772925763, + "grad_norm": 0.011597222648561, + "learning_rate": 0.0017328123476047994, + "loss": 4.166989326477051, + "step": 1687 + }, + { + "epoch": 23.4471615720524, + "grad_norm": 0.01055675558745861, + "learning_rate": 0.001731493743170824, + "loss": 4.172226428985596, + "step": 1688 + }, + { + "epoch": 23.46113537117904, + "grad_norm": 0.011600397527217865, + "learning_rate": 0.0017301749643117506, + "loss": 4.1656951904296875, + "step": 1689 + }, + { + "epoch": 23.475109170305675, + "grad_norm": 0.01114651933312416, + "learning_rate": 0.0017288560120900943, + "loss": 4.189825057983398, + "step": 1690 + }, + { + "epoch": 23.489082969432314, + "grad_norm": 0.009811583906412125, + "learning_rate": 0.0017275368875685126, + "loss": 4.181164741516113, + "step": 1691 + }, + { + "epoch": 23.503056768558952, + "grad_norm": 0.009758968837559223, + "learning_rate": 0.0017262175918097992, + "loss": 4.163188934326172, + "step": 1692 + }, + { + "epoch": 23.51703056768559, + "grad_norm": 0.00999707542359829, + "learning_rate": 0.001724898125876887, + "loss": 4.184117317199707, + "step": 1693 + }, + { + "epoch": 23.531004366812226, + "grad_norm": 0.009528180584311485, + "learning_rate": 0.001723578490832846, + "loss": 4.16407585144043, + "step": 1694 + }, + { + "epoch": 23.544978165938865, + "grad_norm": 0.009717358276247978, + "learning_rate": 0.0017222586877408826, + "loss": 4.209892272949219, + "step": 1695 + }, + { + "epoch": 23.558951965065503, + "grad_norm": 0.010909050703048706, + "learning_rate": 0.0017209387176643377, + "loss": 4.19977331161499, + "step": 1696 + }, + { + "epoch": 23.57292576419214, + "grad_norm": 0.011234750039875507, + "learning_rate": 0.0017196185816666875, + "loss": 4.19363260269165, + "step": 1697 + }, + { + "epoch": 23.586899563318777, + "grad_norm": 0.010300721041858196, + "learning_rate": 0.0017182982808115425, + "loss": 4.142427921295166, + "step": 1698 + }, + { + "epoch": 23.600873362445416, + "grad_norm": 0.00941646471619606, + "learning_rate": 0.001716977816162644, + "loss": 4.159372329711914, + "step": 1699 + }, + { + "epoch": 23.61484716157205, + "grad_norm": 0.009304054081439972, + "learning_rate": 0.0017156571887838679, + "loss": 4.207922458648682, + "step": 1700 + }, + { + "epoch": 23.62882096069869, + "grad_norm": 0.009496180340647697, + "learning_rate": 0.0017143363997392184, + "loss": 4.137894630432129, + "step": 1701 + }, + { + "epoch": 23.64279475982533, + "grad_norm": 0.009257436729967594, + "learning_rate": 0.0017130154500928326, + "loss": 4.198753833770752, + "step": 1702 + }, + { + "epoch": 23.656768558951963, + "grad_norm": 0.00900586973875761, + "learning_rate": 0.0017116943409089746, + "loss": 4.160684585571289, + "step": 1703 + }, + { + "epoch": 23.670742358078602, + "grad_norm": 0.009398494847118855, + "learning_rate": 0.00171037307325204, + "loss": 4.1819353103637695, + "step": 1704 + }, + { + "epoch": 23.68471615720524, + "grad_norm": 0.008633013814687729, + "learning_rate": 0.0017090516481865483, + "loss": 4.176833152770996, + "step": 1705 + }, + { + "epoch": 23.69868995633188, + "grad_norm": 0.00898425467312336, + "learning_rate": 0.0017077300667771484, + "loss": 4.190657138824463, + "step": 1706 + }, + { + "epoch": 23.712663755458514, + "grad_norm": 0.008182398043572903, + "learning_rate": 0.0017064083300886155, + "loss": 4.173429489135742, + "step": 1707 + }, + { + "epoch": 23.726637554585153, + "grad_norm": 0.009589143097400665, + "learning_rate": 0.0017050864391858478, + "loss": 4.1559014320373535, + "step": 1708 + }, + { + "epoch": 23.74061135371179, + "grad_norm": 0.010304383002221584, + "learning_rate": 0.0017037643951338698, + "loss": 4.155453681945801, + "step": 1709 + }, + { + "epoch": 23.754585152838427, + "grad_norm": 0.009943748824298382, + "learning_rate": 0.0017024421989978275, + "loss": 4.119734287261963, + "step": 1710 + }, + { + "epoch": 23.768558951965066, + "grad_norm": 0.009238033555448055, + "learning_rate": 0.001701119851842992, + "loss": 4.188974857330322, + "step": 1711 + }, + { + "epoch": 23.782532751091704, + "grad_norm": 0.008058690465986729, + "learning_rate": 0.0016997973547347531, + "loss": 4.19934606552124, + "step": 1712 + }, + { + "epoch": 23.79650655021834, + "grad_norm": 0.008485953323543072, + "learning_rate": 0.0016984747087386233, + "loss": 4.165615558624268, + "step": 1713 + }, + { + "epoch": 23.810480349344978, + "grad_norm": 0.008476605638861656, + "learning_rate": 0.0016971519149202347, + "loss": 4.21021842956543, + "step": 1714 + }, + { + "epoch": 23.824454148471617, + "grad_norm": 0.008588708005845547, + "learning_rate": 0.0016958289743453385, + "loss": 4.208064556121826, + "step": 1715 + }, + { + "epoch": 23.83842794759825, + "grad_norm": 0.009223178029060364, + "learning_rate": 0.0016945058880798037, + "loss": 4.158413887023926, + "step": 1716 + }, + { + "epoch": 23.85240174672489, + "grad_norm": 0.009196167811751366, + "learning_rate": 0.0016931826571896166, + "loss": 4.203187465667725, + "step": 1717 + }, + { + "epoch": 23.86637554585153, + "grad_norm": 0.009488921612501144, + "learning_rate": 0.001691859282740882, + "loss": 4.178227424621582, + "step": 1718 + }, + { + "epoch": 23.880349344978168, + "grad_norm": 0.009253161028027534, + "learning_rate": 0.0016905357657998167, + "loss": 4.131230354309082, + "step": 1719 + }, + { + "epoch": 23.894323144104803, + "grad_norm": 0.010069048963487148, + "learning_rate": 0.0016892121074327559, + "loss": 4.210734844207764, + "step": 1720 + }, + { + "epoch": 23.90829694323144, + "grad_norm": 0.009945095516741276, + "learning_rate": 0.001687888308706147, + "loss": 4.184042930603027, + "step": 1721 + }, + { + "epoch": 23.92227074235808, + "grad_norm": 0.010483640246093273, + "learning_rate": 0.0016865643706865505, + "loss": 4.156919479370117, + "step": 1722 + }, + { + "epoch": 23.936244541484715, + "grad_norm": 0.011560715734958649, + "learning_rate": 0.0016852402944406391, + "loss": 4.185520172119141, + "step": 1723 + }, + { + "epoch": 23.950218340611354, + "grad_norm": 0.010001277551054955, + "learning_rate": 0.0016839160810351978, + "loss": 4.226828575134277, + "step": 1724 + }, + { + "epoch": 23.964192139737992, + "grad_norm": 0.010335934348404408, + "learning_rate": 0.0016825917315371202, + "loss": 4.183531284332275, + "step": 1725 + }, + { + "epoch": 23.978165938864628, + "grad_norm": 0.011322776786983013, + "learning_rate": 0.001681267247013412, + "loss": 4.179635047912598, + "step": 1726 + }, + { + "epoch": 23.992139737991266, + "grad_norm": 0.009663194417953491, + "learning_rate": 0.001679942628531186, + "loss": 4.165654182434082, + "step": 1727 + }, + { + "epoch": 24.0, + "grad_norm": 0.010590246878564358, + "learning_rate": 0.0016786178771576643, + "loss": 4.1539459228515625, + "step": 1728 + }, + { + "epoch": 24.0, + "eval_loss": 4.511386394500732, + "eval_runtime": 51.7751, + "eval_samples_per_second": 47.166, + "eval_steps_per_second": 1.487, + "step": 1728 + }, + { + "epoch": 24.01397379912664, + "grad_norm": 0.01254354789853096, + "learning_rate": 0.0016772929939601728, + "loss": 4.104002475738525, + "step": 1729 + }, + { + "epoch": 24.027947598253274, + "grad_norm": 0.01300987508147955, + "learning_rate": 0.0016759679800061483, + "loss": 4.127395153045654, + "step": 1730 + }, + { + "epoch": 24.041921397379912, + "grad_norm": 0.013325816951692104, + "learning_rate": 0.0016746428363631299, + "loss": 4.131229400634766, + "step": 1731 + }, + { + "epoch": 24.05589519650655, + "grad_norm": 0.014302870258688927, + "learning_rate": 0.001673317564098761, + "loss": 4.079358100891113, + "step": 1732 + }, + { + "epoch": 24.069868995633186, + "grad_norm": 0.013776068575680256, + "learning_rate": 0.0016719921642807905, + "loss": 4.132065773010254, + "step": 1733 + }, + { + "epoch": 24.083842794759825, + "grad_norm": 0.010788560844957829, + "learning_rate": 0.0016706666379770689, + "loss": 4.134626388549805, + "step": 1734 + }, + { + "epoch": 24.097816593886463, + "grad_norm": 0.009766630828380585, + "learning_rate": 0.0016693409862555493, + "loss": 4.1645426750183105, + "step": 1735 + }, + { + "epoch": 24.111790393013102, + "grad_norm": 0.011760029010474682, + "learning_rate": 0.0016680152101842842, + "loss": 4.110664367675781, + "step": 1736 + }, + { + "epoch": 24.125764192139737, + "grad_norm": 0.011047836393117905, + "learning_rate": 0.0016666893108314288, + "loss": 4.129222869873047, + "step": 1737 + }, + { + "epoch": 24.139737991266376, + "grad_norm": 0.009989139623939991, + "learning_rate": 0.0016653632892652344, + "loss": 4.076835632324219, + "step": 1738 + }, + { + "epoch": 24.153711790393015, + "grad_norm": 0.010033783502876759, + "learning_rate": 0.001664037146554054, + "loss": 4.11555290222168, + "step": 1739 + }, + { + "epoch": 24.16768558951965, + "grad_norm": 0.00996427796781063, + "learning_rate": 0.001662710883766337, + "loss": 4.131463527679443, + "step": 1740 + }, + { + "epoch": 24.18165938864629, + "grad_norm": 0.01044023409485817, + "learning_rate": 0.0016613845019706288, + "loss": 4.098750114440918, + "step": 1741 + }, + { + "epoch": 24.195633187772927, + "grad_norm": 0.010033872909843922, + "learning_rate": 0.0016600580022355712, + "loss": 4.161447048187256, + "step": 1742 + }, + { + "epoch": 24.209606986899562, + "grad_norm": 0.009329567663371563, + "learning_rate": 0.0016587313856299012, + "loss": 4.146491527557373, + "step": 1743 + }, + { + "epoch": 24.2235807860262, + "grad_norm": 0.010061589069664478, + "learning_rate": 0.0016574046532224504, + "loss": 4.107517719268799, + "step": 1744 + }, + { + "epoch": 24.23755458515284, + "grad_norm": 0.009491910226643085, + "learning_rate": 0.0016560778060821424, + "loss": 4.098226547241211, + "step": 1745 + }, + { + "epoch": 24.251528384279474, + "grad_norm": 0.008529345504939556, + "learning_rate": 0.0016547508452779943, + "loss": 4.1247711181640625, + "step": 1746 + }, + { + "epoch": 24.265502183406113, + "grad_norm": 0.00936342403292656, + "learning_rate": 0.0016534237718791136, + "loss": 4.120573043823242, + "step": 1747 + }, + { + "epoch": 24.27947598253275, + "grad_norm": 0.010255641303956509, + "learning_rate": 0.0016520965869547008, + "loss": 4.089067459106445, + "step": 1748 + }, + { + "epoch": 24.29344978165939, + "grad_norm": 0.011638758704066277, + "learning_rate": 0.001650769291574044, + "loss": 4.074847221374512, + "step": 1749 + }, + { + "epoch": 24.307423580786025, + "grad_norm": 0.014707857742905617, + "learning_rate": 0.0016494418868065214, + "loss": 4.153377532958984, + "step": 1750 + }, + { + "epoch": 24.321397379912664, + "grad_norm": 0.01650111749768257, + "learning_rate": 0.0016481143737215986, + "loss": 4.134062767028809, + "step": 1751 + }, + { + "epoch": 24.335371179039303, + "grad_norm": 0.014833372086286545, + "learning_rate": 0.00164678675338883, + "loss": 4.1344828605651855, + "step": 1752 + }, + { + "epoch": 24.349344978165938, + "grad_norm": 0.012257886119186878, + "learning_rate": 0.0016454590268778538, + "loss": 4.098813533782959, + "step": 1753 + }, + { + "epoch": 24.363318777292577, + "grad_norm": 0.013508322648704052, + "learning_rate": 0.0016441311952583959, + "loss": 4.106297492980957, + "step": 1754 + }, + { + "epoch": 24.377292576419215, + "grad_norm": 0.01256900280714035, + "learning_rate": 0.001642803259600267, + "loss": 4.143631935119629, + "step": 1755 + }, + { + "epoch": 24.39126637554585, + "grad_norm": 0.009885406121611595, + "learning_rate": 0.00164147522097336, + "loss": 4.120885848999023, + "step": 1756 + }, + { + "epoch": 24.40524017467249, + "grad_norm": 0.012004674412310123, + "learning_rate": 0.0016401470804476518, + "loss": 4.089243412017822, + "step": 1757 + }, + { + "epoch": 24.419213973799128, + "grad_norm": 0.01224028691649437, + "learning_rate": 0.0016388188390932012, + "loss": 4.077765464782715, + "step": 1758 + }, + { + "epoch": 24.433187772925763, + "grad_norm": 0.012511290609836578, + "learning_rate": 0.001637490497980149, + "loss": 4.161346435546875, + "step": 1759 + }, + { + "epoch": 24.4471615720524, + "grad_norm": 0.010701716877520084, + "learning_rate": 0.0016361620581787143, + "loss": 4.10698938369751, + "step": 1760 + }, + { + "epoch": 24.46113537117904, + "grad_norm": 0.010713113471865654, + "learning_rate": 0.001634833520759198, + "loss": 4.114283561706543, + "step": 1761 + }, + { + "epoch": 24.475109170305675, + "grad_norm": 0.011018907651305199, + "learning_rate": 0.001633504886791978, + "loss": 4.134284019470215, + "step": 1762 + }, + { + "epoch": 24.489082969432314, + "grad_norm": 0.011279584839940071, + "learning_rate": 0.0016321761573475117, + "loss": 4.1456804275512695, + "step": 1763 + }, + { + "epoch": 24.503056768558952, + "grad_norm": 0.010642931796610355, + "learning_rate": 0.001630847333496332, + "loss": 4.160918235778809, + "step": 1764 + }, + { + "epoch": 24.51703056768559, + "grad_norm": 0.011069678701460361, + "learning_rate": 0.0016295184163090482, + "loss": 4.136838912963867, + "step": 1765 + }, + { + "epoch": 24.531004366812226, + "grad_norm": 0.011007166467607021, + "learning_rate": 0.0016281894068563443, + "loss": 4.129153728485107, + "step": 1766 + }, + { + "epoch": 24.544978165938865, + "grad_norm": 0.010055703110992908, + "learning_rate": 0.0016268603062089802, + "loss": 4.138665676116943, + "step": 1767 + }, + { + "epoch": 24.558951965065503, + "grad_norm": 0.009839230217039585, + "learning_rate": 0.0016255311154377875, + "loss": 4.1366496086120605, + "step": 1768 + }, + { + "epoch": 24.57292576419214, + "grad_norm": 0.011191432364284992, + "learning_rate": 0.0016242018356136714, + "loss": 4.129880905151367, + "step": 1769 + }, + { + "epoch": 24.586899563318777, + "grad_norm": 0.012130341492593288, + "learning_rate": 0.0016228724678076095, + "loss": 4.1425981521606445, + "step": 1770 + }, + { + "epoch": 24.600873362445416, + "grad_norm": 0.012707795947790146, + "learning_rate": 0.0016215430130906482, + "loss": 4.115452766418457, + "step": 1771 + }, + { + "epoch": 24.61484716157205, + "grad_norm": 0.012707809917628765, + "learning_rate": 0.001620213472533906, + "loss": 4.122758865356445, + "step": 1772 + }, + { + "epoch": 24.62882096069869, + "grad_norm": 0.01182397548109293, + "learning_rate": 0.001618883847208569, + "loss": 4.140275955200195, + "step": 1773 + }, + { + "epoch": 24.64279475982533, + "grad_norm": 0.010865340940654278, + "learning_rate": 0.001617554138185893, + "loss": 4.137759208679199, + "step": 1774 + }, + { + "epoch": 24.656768558951963, + "grad_norm": 0.010973592288792133, + "learning_rate": 0.0016162243465371998, + "loss": 4.138947010040283, + "step": 1775 + }, + { + "epoch": 24.670742358078602, + "grad_norm": 0.011247226968407631, + "learning_rate": 0.001614894473333879, + "loss": 4.115516662597656, + "step": 1776 + }, + { + "epoch": 24.68471615720524, + "grad_norm": 0.010837487876415253, + "learning_rate": 0.0016135645196473853, + "loss": 4.160427093505859, + "step": 1777 + }, + { + "epoch": 24.69868995633188, + "grad_norm": 0.009096396155655384, + "learning_rate": 0.0016122344865492385, + "loss": 4.1200032234191895, + "step": 1778 + }, + { + "epoch": 24.712663755458514, + "grad_norm": 0.010671247728168964, + "learning_rate": 0.0016109043751110223, + "loss": 4.097919464111328, + "step": 1779 + }, + { + "epoch": 24.726637554585153, + "grad_norm": 0.0107562355697155, + "learning_rate": 0.0016095741864043831, + "loss": 4.096290588378906, + "step": 1780 + }, + { + "epoch": 24.74061135371179, + "grad_norm": 0.008909327909350395, + "learning_rate": 0.0016082439215010305, + "loss": 4.171017646789551, + "step": 1781 + }, + { + "epoch": 24.754585152838427, + "grad_norm": 0.010305124334990978, + "learning_rate": 0.0016069135814727343, + "loss": 4.1450276374816895, + "step": 1782 + }, + { + "epoch": 24.768558951965066, + "grad_norm": 0.011235613375902176, + "learning_rate": 0.0016055831673913255, + "loss": 4.134704113006592, + "step": 1783 + }, + { + "epoch": 24.782532751091704, + "grad_norm": 0.00825632456690073, + "learning_rate": 0.0016042526803286949, + "loss": 4.107443332672119, + "step": 1784 + }, + { + "epoch": 24.79650655021834, + "grad_norm": 0.009056378155946732, + "learning_rate": 0.0016029221213567921, + "loss": 4.132656097412109, + "step": 1785 + }, + { + "epoch": 24.810480349344978, + "grad_norm": 0.008835164830088615, + "learning_rate": 0.001601591491547624, + "loss": 4.124729633331299, + "step": 1786 + }, + { + "epoch": 24.824454148471617, + "grad_norm": 0.00915333442389965, + "learning_rate": 0.0016002607919732557, + "loss": 4.119500160217285, + "step": 1787 + }, + { + "epoch": 24.83842794759825, + "grad_norm": 0.008532539941370487, + "learning_rate": 0.0015989300237058066, + "loss": 4.141661643981934, + "step": 1788 + }, + { + "epoch": 24.85240174672489, + "grad_norm": 0.009952311404049397, + "learning_rate": 0.0015975991878174533, + "loss": 4.140929222106934, + "step": 1789 + }, + { + "epoch": 24.86637554585153, + "grad_norm": 0.008920297026634216, + "learning_rate": 0.0015962682853804268, + "loss": 4.155877113342285, + "step": 1790 + }, + { + "epoch": 24.880349344978168, + "grad_norm": 0.008770049549639225, + "learning_rate": 0.0015949373174670104, + "loss": 4.126394271850586, + "step": 1791 + }, + { + "epoch": 24.894323144104803, + "grad_norm": 0.0078084105625748634, + "learning_rate": 0.0015936062851495411, + "loss": 4.128876209259033, + "step": 1792 + }, + { + "epoch": 24.90829694323144, + "grad_norm": 0.00816641841083765, + "learning_rate": 0.0015922751895004078, + "loss": 4.137410640716553, + "step": 1793 + }, + { + "epoch": 24.92227074235808, + "grad_norm": 0.008895459584891796, + "learning_rate": 0.0015909440315920505, + "loss": 4.07957649230957, + "step": 1794 + }, + { + "epoch": 24.936244541484715, + "grad_norm": 0.008551528677344322, + "learning_rate": 0.001589612812496958, + "loss": 4.104213714599609, + "step": 1795 + }, + { + "epoch": 24.950218340611354, + "grad_norm": 0.008325488306581974, + "learning_rate": 0.0015882815332876707, + "loss": 4.1584792137146, + "step": 1796 + }, + { + "epoch": 24.964192139737992, + "grad_norm": 0.009341313503682613, + "learning_rate": 0.0015869501950367744, + "loss": 4.1417436599731445, + "step": 1797 + }, + { + "epoch": 24.978165938864628, + "grad_norm": 0.00999050959944725, + "learning_rate": 0.0015856187988169067, + "loss": 4.136049270629883, + "step": 1798 + }, + { + "epoch": 24.992139737991266, + "grad_norm": 0.008889932185411453, + "learning_rate": 0.0015842873457007484, + "loss": 4.16257381439209, + "step": 1799 + }, + { + "epoch": 25.0, + "grad_norm": 0.009109573438763618, + "learning_rate": 0.0015829558367610268, + "loss": 4.171441555023193, + "step": 1800 + }, + { + "epoch": 25.0, + "eval_loss": 4.485942840576172, + "eval_runtime": 52.1051, + "eval_samples_per_second": 46.867, + "eval_steps_per_second": 1.478, + "step": 1800 + }, + { + "epoch": 25.01397379912664, + "grad_norm": 0.009881424717605114, + "learning_rate": 0.001581624273070515, + "loss": 4.059608459472656, + "step": 1801 + }, + { + "epoch": 25.027947598253274, + "grad_norm": 0.011156721040606499, + "learning_rate": 0.0015802926557020294, + "loss": 4.047014236450195, + "step": 1802 + }, + { + "epoch": 25.041921397379912, + "grad_norm": 0.012326992116868496, + "learning_rate": 0.0015789609857284309, + "loss": 4.06541109085083, + "step": 1803 + }, + { + "epoch": 25.05589519650655, + "grad_norm": 0.01332208700478077, + "learning_rate": 0.0015776292642226213, + "loss": 4.042714595794678, + "step": 1804 + }, + { + "epoch": 25.069868995633186, + "grad_norm": 0.012720128521323204, + "learning_rate": 0.0015762974922575446, + "loss": 4.098389625549316, + "step": 1805 + }, + { + "epoch": 25.083842794759825, + "grad_norm": 0.010723215527832508, + "learning_rate": 0.0015749656709061854, + "loss": 4.059185981750488, + "step": 1806 + }, + { + "epoch": 25.097816593886463, + "grad_norm": 0.010605281218886375, + "learning_rate": 0.001573633801241569, + "loss": 4.099538326263428, + "step": 1807 + }, + { + "epoch": 25.111790393013102, + "grad_norm": 0.01101530622690916, + "learning_rate": 0.0015723018843367576, + "loss": 4.095674514770508, + "step": 1808 + }, + { + "epoch": 25.125764192139737, + "grad_norm": 0.01135838869959116, + "learning_rate": 0.0015709699212648532, + "loss": 4.070535659790039, + "step": 1809 + }, + { + "epoch": 25.139737991266376, + "grad_norm": 0.012119271792471409, + "learning_rate": 0.0015696379130989936, + "loss": 4.060605049133301, + "step": 1810 + }, + { + "epoch": 25.153711790393015, + "grad_norm": 0.012783961370587349, + "learning_rate": 0.001568305860912355, + "loss": 4.075931072235107, + "step": 1811 + }, + { + "epoch": 25.16768558951965, + "grad_norm": 0.011486220173537731, + "learning_rate": 0.0015669737657781461, + "loss": 4.075855255126953, + "step": 1812 + }, + { + "epoch": 25.18165938864629, + "grad_norm": 0.01037552859634161, + "learning_rate": 0.0015656416287696136, + "loss": 4.087729454040527, + "step": 1813 + }, + { + "epoch": 25.195633187772927, + "grad_norm": 0.009802743792533875, + "learning_rate": 0.0015643094509600354, + "loss": 4.0855607986450195, + "step": 1814 + }, + { + "epoch": 25.209606986899562, + "grad_norm": 0.010804438032209873, + "learning_rate": 0.001562977233422723, + "loss": 4.064528465270996, + "step": 1815 + }, + { + "epoch": 25.2235807860262, + "grad_norm": 0.009505728259682655, + "learning_rate": 0.0015616449772310206, + "loss": 4.036775588989258, + "step": 1816 + }, + { + "epoch": 25.23755458515284, + "grad_norm": 0.010634485632181168, + "learning_rate": 0.0015603126834583024, + "loss": 4.069923400878906, + "step": 1817 + }, + { + "epoch": 25.251528384279474, + "grad_norm": 0.011392096057534218, + "learning_rate": 0.0015589803531779736, + "loss": 4.0812177658081055, + "step": 1818 + }, + { + "epoch": 25.265502183406113, + "grad_norm": 0.011851722374558449, + "learning_rate": 0.0015576479874634683, + "loss": 4.09183406829834, + "step": 1819 + }, + { + "epoch": 25.27947598253275, + "grad_norm": 0.011205444112420082, + "learning_rate": 0.0015563155873882502, + "loss": 4.07825231552124, + "step": 1820 + }, + { + "epoch": 25.29344978165939, + "grad_norm": 0.009676349349319935, + "learning_rate": 0.0015549831540258093, + "loss": 4.052826404571533, + "step": 1821 + }, + { + "epoch": 25.307423580786025, + "grad_norm": 0.009376165457069874, + "learning_rate": 0.0015536506884496638, + "loss": 4.02100133895874, + "step": 1822 + }, + { + "epoch": 25.321397379912664, + "grad_norm": 0.00936955213546753, + "learning_rate": 0.0015523181917333564, + "loss": 4.049526214599609, + "step": 1823 + }, + { + "epoch": 25.335371179039303, + "grad_norm": 0.00937784556299448, + "learning_rate": 0.0015509856649504564, + "loss": 4.096356391906738, + "step": 1824 + }, + { + "epoch": 25.349344978165938, + "grad_norm": 0.010014008730649948, + "learning_rate": 0.0015496531091745555, + "loss": 4.03254508972168, + "step": 1825 + }, + { + "epoch": 25.363318777292577, + "grad_norm": 0.009909634478390217, + "learning_rate": 0.0015483205254792705, + "loss": 4.067630767822266, + "step": 1826 + }, + { + "epoch": 25.377292576419215, + "grad_norm": 0.009319501928985119, + "learning_rate": 0.0015469879149382399, + "loss": 4.094167709350586, + "step": 1827 + }, + { + "epoch": 25.39126637554585, + "grad_norm": 0.008634721860289574, + "learning_rate": 0.0015456552786251244, + "loss": 4.122404098510742, + "step": 1828 + }, + { + "epoch": 25.40524017467249, + "grad_norm": 0.009412454441189766, + "learning_rate": 0.0015443226176136049, + "loss": 4.081546783447266, + "step": 1829 + }, + { + "epoch": 25.419213973799128, + "grad_norm": 0.010102108120918274, + "learning_rate": 0.0015429899329773814, + "loss": 4.049839019775391, + "step": 1830 + }, + { + "epoch": 25.433187772925763, + "grad_norm": 0.010921654291450977, + "learning_rate": 0.0015416572257901746, + "loss": 4.097097396850586, + "step": 1831 + }, + { + "epoch": 25.4471615720524, + "grad_norm": 0.010607713833451271, + "learning_rate": 0.0015403244971257226, + "loss": 4.084481239318848, + "step": 1832 + }, + { + "epoch": 25.46113537117904, + "grad_norm": 0.011887960135936737, + "learning_rate": 0.0015389917480577806, + "loss": 4.074968338012695, + "step": 1833 + }, + { + "epoch": 25.475109170305675, + "grad_norm": 0.01089888159185648, + "learning_rate": 0.0015376589796601203, + "loss": 4.062132835388184, + "step": 1834 + }, + { + "epoch": 25.489082969432314, + "grad_norm": 0.009934183210134506, + "learning_rate": 0.0015363261930065298, + "loss": 4.128092288970947, + "step": 1835 + }, + { + "epoch": 25.503056768558952, + "grad_norm": 0.00878534372895956, + "learning_rate": 0.0015349933891708105, + "loss": 4.069051742553711, + "step": 1836 + }, + { + "epoch": 25.51703056768559, + "grad_norm": 0.009928155690431595, + "learning_rate": 0.0015336605692267788, + "loss": 4.073912620544434, + "step": 1837 + }, + { + "epoch": 25.531004366812226, + "grad_norm": 0.011091411113739014, + "learning_rate": 0.0015323277342482637, + "loss": 4.125944137573242, + "step": 1838 + }, + { + "epoch": 25.544978165938865, + "grad_norm": 0.010314172133803368, + "learning_rate": 0.0015309948853091055, + "loss": 4.062620162963867, + "step": 1839 + }, + { + "epoch": 25.558951965065503, + "grad_norm": 0.010423967614769936, + "learning_rate": 0.001529662023483158, + "loss": 4.101737976074219, + "step": 1840 + }, + { + "epoch": 25.57292576419214, + "grad_norm": 0.009822997264564037, + "learning_rate": 0.0015283291498442817, + "loss": 4.064520359039307, + "step": 1841 + }, + { + "epoch": 25.586899563318777, + "grad_norm": 0.009757095016539097, + "learning_rate": 0.001526996265466351, + "loss": 4.100624084472656, + "step": 1842 + }, + { + "epoch": 25.600873362445416, + "grad_norm": 0.008650688454508781, + "learning_rate": 0.001525663371423246, + "loss": 4.109685897827148, + "step": 1843 + }, + { + "epoch": 25.61484716157205, + "grad_norm": 0.007378234062343836, + "learning_rate": 0.0015243304687888553, + "loss": 4.044726848602295, + "step": 1844 + }, + { + "epoch": 25.62882096069869, + "grad_norm": 0.008194336667656898, + "learning_rate": 0.001522997558637074, + "loss": 4.097923278808594, + "step": 1845 + }, + { + "epoch": 25.64279475982533, + "grad_norm": 0.009764892049133778, + "learning_rate": 0.0015216646420418048, + "loss": 4.102148056030273, + "step": 1846 + }, + { + "epoch": 25.656768558951963, + "grad_norm": 0.009242868050932884, + "learning_rate": 0.001520331720076953, + "loss": 4.068563461303711, + "step": 1847 + }, + { + "epoch": 25.670742358078602, + "grad_norm": 0.010409324429929256, + "learning_rate": 0.001518998793816431, + "loss": 4.065944194793701, + "step": 1848 + }, + { + "epoch": 25.68471615720524, + "grad_norm": 0.009215258061885834, + "learning_rate": 0.0015176658643341527, + "loss": 4.10223388671875, + "step": 1849 + }, + { + "epoch": 25.69868995633188, + "grad_norm": 0.008065393194556236, + "learning_rate": 0.0015163329327040366, + "loss": 4.099432945251465, + "step": 1850 + }, + { + "epoch": 25.712663755458514, + "grad_norm": 0.008498288691043854, + "learning_rate": 0.001515, + "loss": 4.109854698181152, + "step": 1851 + }, + { + "epoch": 25.726637554585153, + "grad_norm": 0.008919150568544865, + "learning_rate": 0.0015136670672959633, + "loss": 4.128247261047363, + "step": 1852 + }, + { + "epoch": 25.74061135371179, + "grad_norm": 0.009104429744184017, + "learning_rate": 0.0015123341356658474, + "loss": 4.068622589111328, + "step": 1853 + }, + { + "epoch": 25.754585152838427, + "grad_norm": 0.008558365516364574, + "learning_rate": 0.0015110012061835692, + "loss": 4.114214897155762, + "step": 1854 + }, + { + "epoch": 25.768558951965066, + "grad_norm": 0.007865076884627342, + "learning_rate": 0.0015096682799230477, + "loss": 4.11155366897583, + "step": 1855 + }, + { + "epoch": 25.782532751091704, + "grad_norm": 0.007848534733057022, + "learning_rate": 0.0015083353579581958, + "loss": 4.045694351196289, + "step": 1856 + }, + { + "epoch": 25.79650655021834, + "grad_norm": 0.008238030597567558, + "learning_rate": 0.0015070024413629264, + "loss": 4.093930721282959, + "step": 1857 + }, + { + "epoch": 25.810480349344978, + "grad_norm": 0.007536398712545633, + "learning_rate": 0.0015056695312111448, + "loss": 4.106235504150391, + "step": 1858 + }, + { + "epoch": 25.824454148471617, + "grad_norm": 0.008156892843544483, + "learning_rate": 0.001504336628576754, + "loss": 4.13060998916626, + "step": 1859 + }, + { + "epoch": 25.83842794759825, + "grad_norm": 0.008144439198076725, + "learning_rate": 0.0015030037345336487, + "loss": 4.091304779052734, + "step": 1860 + }, + { + "epoch": 25.85240174672489, + "grad_norm": 0.007757358253002167, + "learning_rate": 0.001501670850155718, + "loss": 4.069276809692383, + "step": 1861 + }, + { + "epoch": 25.86637554585153, + "grad_norm": 0.00860623549669981, + "learning_rate": 0.0015003379765168425, + "loss": 4.107658386230469, + "step": 1862 + }, + { + "epoch": 25.880349344978168, + "grad_norm": 0.008614457212388515, + "learning_rate": 0.0014990051146908949, + "loss": 4.106705665588379, + "step": 1863 + }, + { + "epoch": 25.894323144104803, + "grad_norm": 0.009120592847466469, + "learning_rate": 0.001497672265751737, + "loss": 4.0891923904418945, + "step": 1864 + }, + { + "epoch": 25.90829694323144, + "grad_norm": 0.009446980431675911, + "learning_rate": 0.0014963394307732213, + "loss": 4.098847389221191, + "step": 1865 + }, + { + "epoch": 25.92227074235808, + "grad_norm": 0.010494766756892204, + "learning_rate": 0.0014950066108291894, + "loss": 4.115325450897217, + "step": 1866 + }, + { + "epoch": 25.936244541484715, + "grad_norm": 0.010057658888399601, + "learning_rate": 0.0014936738069934706, + "loss": 4.099984169006348, + "step": 1867 + }, + { + "epoch": 25.950218340611354, + "grad_norm": 0.009071771055459976, + "learning_rate": 0.0014923410203398796, + "loss": 4.091543674468994, + "step": 1868 + }, + { + "epoch": 25.964192139737992, + "grad_norm": 0.009937075898051262, + "learning_rate": 0.00149100825194222, + "loss": 4.086492538452148, + "step": 1869 + }, + { + "epoch": 25.978165938864628, + "grad_norm": 0.010222986340522766, + "learning_rate": 0.0014896755028742778, + "loss": 4.063841819763184, + "step": 1870 + }, + { + "epoch": 25.992139737991266, + "grad_norm": 0.010513762012124062, + "learning_rate": 0.001488342774209826, + "loss": 4.11991548538208, + "step": 1871 + }, + { + "epoch": 26.0, + "grad_norm": 0.009530222974717617, + "learning_rate": 0.0014870100670226192, + "loss": 4.043334007263184, + "step": 1872 + }, + { + "epoch": 26.0, + "eval_loss": 4.467459678649902, + "eval_runtime": 54.1628, + "eval_samples_per_second": 45.086, + "eval_steps_per_second": 1.422, + "step": 1872 + }, + { + "epoch": 26.01397379912664, + "grad_norm": 0.010274187661707401, + "learning_rate": 0.0014856773823863955, + "loss": 4.049561500549316, + "step": 1873 + }, + { + "epoch": 26.027947598253274, + "grad_norm": 0.011448442004621029, + "learning_rate": 0.001484344721374876, + "loss": 4.019484996795654, + "step": 1874 + }, + { + "epoch": 26.041921397379912, + "grad_norm": 0.011361192911863327, + "learning_rate": 0.0014830120850617599, + "loss": 4.0429887771606445, + "step": 1875 + }, + { + "epoch": 26.05589519650655, + "grad_norm": 0.010250156745314598, + "learning_rate": 0.0014816794745207299, + "loss": 4.012682914733887, + "step": 1876 + }, + { + "epoch": 26.069868995633186, + "grad_norm": 0.010024400427937508, + "learning_rate": 0.0014803468908254449, + "loss": 3.979409694671631, + "step": 1877 + }, + { + "epoch": 26.083842794759825, + "grad_norm": 0.01153535582125187, + "learning_rate": 0.0014790143350495446, + "loss": 4.045780181884766, + "step": 1878 + }, + { + "epoch": 26.097816593886463, + "grad_norm": 0.012632817961275578, + "learning_rate": 0.001477681808266644, + "loss": 4.057641983032227, + "step": 1879 + }, + { + "epoch": 26.111790393013102, + "grad_norm": 0.013758293353021145, + "learning_rate": 0.0014763493115503366, + "loss": 4.061607360839844, + "step": 1880 + }, + { + "epoch": 26.125764192139737, + "grad_norm": 0.014041583053767681, + "learning_rate": 0.001475016845974191, + "loss": 4.035029411315918, + "step": 1881 + }, + { + "epoch": 26.139737991266376, + "grad_norm": 0.013629456982016563, + "learning_rate": 0.0014736844126117497, + "loss": 4.062169075012207, + "step": 1882 + }, + { + "epoch": 26.153711790393015, + "grad_norm": 0.011023213155567646, + "learning_rate": 0.0014723520125365323, + "loss": 4.02731990814209, + "step": 1883 + }, + { + "epoch": 26.16768558951965, + "grad_norm": 0.01001750584691763, + "learning_rate": 0.0014710196468220266, + "loss": 4.019882678985596, + "step": 1884 + }, + { + "epoch": 26.18165938864629, + "grad_norm": 0.008321646600961685, + "learning_rate": 0.001469687316541698, + "loss": 4.005446434020996, + "step": 1885 + }, + { + "epoch": 26.195633187772927, + "grad_norm": 0.008935946971178055, + "learning_rate": 0.0014683550227689796, + "loss": 4.017102241516113, + "step": 1886 + }, + { + "epoch": 26.209606986899562, + "grad_norm": 0.009137130342423916, + "learning_rate": 0.001467022766577277, + "loss": 4.024102210998535, + "step": 1887 + }, + { + "epoch": 26.2235807860262, + "grad_norm": 0.00805022194981575, + "learning_rate": 0.001465690549039965, + "loss": 3.9935250282287598, + "step": 1888 + }, + { + "epoch": 26.23755458515284, + "grad_norm": 0.009204058907926083, + "learning_rate": 0.0014643583712303865, + "loss": 4.042818546295166, + "step": 1889 + }, + { + "epoch": 26.251528384279474, + "grad_norm": 0.009812815114855766, + "learning_rate": 0.001463026234221854, + "loss": 4.0252485275268555, + "step": 1890 + }, + { + "epoch": 26.265502183406113, + "grad_norm": 0.009955782443284988, + "learning_rate": 0.0014616941390876454, + "loss": 3.9839768409729004, + "step": 1891 + }, + { + "epoch": 26.27947598253275, + "grad_norm": 0.009374645538628101, + "learning_rate": 0.001460362086901007, + "loss": 4.040245056152344, + "step": 1892 + }, + { + "epoch": 26.29344978165939, + "grad_norm": 0.010008739307522774, + "learning_rate": 0.0014590300787351471, + "loss": 4.0306596755981445, + "step": 1893 + }, + { + "epoch": 26.307423580786025, + "grad_norm": 0.009014200419187546, + "learning_rate": 0.001457698115663243, + "loss": 4.028713226318359, + "step": 1894 + }, + { + "epoch": 26.321397379912664, + "grad_norm": 0.008381523191928864, + "learning_rate": 0.0014563661987584313, + "loss": 4.024718284606934, + "step": 1895 + }, + { + "epoch": 26.335371179039303, + "grad_norm": 0.008627948351204395, + "learning_rate": 0.0014550343290938143, + "loss": 4.02303409576416, + "step": 1896 + }, + { + "epoch": 26.349344978165938, + "grad_norm": 0.010257705114781857, + "learning_rate": 0.0014537025077424556, + "loss": 4.061058044433594, + "step": 1897 + }, + { + "epoch": 26.363318777292577, + "grad_norm": 0.010759209282696247, + "learning_rate": 0.0014523707357773788, + "loss": 4.064718246459961, + "step": 1898 + }, + { + "epoch": 26.377292576419215, + "grad_norm": 0.009966139681637287, + "learning_rate": 0.0014510390142715697, + "loss": 3.9919447898864746, + "step": 1899 + }, + { + "epoch": 26.39126637554585, + "grad_norm": 0.00997584406286478, + "learning_rate": 0.001449707344297971, + "loss": 4.0431599617004395, + "step": 1900 + }, + { + "epoch": 26.40524017467249, + "grad_norm": 0.010846944525837898, + "learning_rate": 0.0014483757269294858, + "loss": 4.0222697257995605, + "step": 1901 + }, + { + "epoch": 26.419213973799128, + "grad_norm": 0.00991352554410696, + "learning_rate": 0.0014470441632389735, + "loss": 4.059438705444336, + "step": 1902 + }, + { + "epoch": 26.433187772925763, + "grad_norm": 0.008994397707283497, + "learning_rate": 0.0014457126542992518, + "loss": 4.021494388580322, + "step": 1903 + }, + { + "epoch": 26.4471615720524, + "grad_norm": 0.009567267261445522, + "learning_rate": 0.0014443812011830934, + "loss": 4.070810317993164, + "step": 1904 + }, + { + "epoch": 26.46113537117904, + "grad_norm": 0.009199011139571667, + "learning_rate": 0.0014430498049632255, + "loss": 4.020938873291016, + "step": 1905 + }, + { + "epoch": 26.475109170305675, + "grad_norm": 0.010440350510179996, + "learning_rate": 0.0014417184667123303, + "loss": 4.062529563903809, + "step": 1906 + }, + { + "epoch": 26.489082969432314, + "grad_norm": 0.009962482377886772, + "learning_rate": 0.0014403871875030424, + "loss": 4.0687150955200195, + "step": 1907 + }, + { + "epoch": 26.503056768558952, + "grad_norm": 0.009389379993081093, + "learning_rate": 0.0014390559684079502, + "loss": 4.055185794830322, + "step": 1908 + }, + { + "epoch": 26.51703056768559, + "grad_norm": 0.011469941586256027, + "learning_rate": 0.0014377248104995924, + "loss": 4.041023254394531, + "step": 1909 + }, + { + "epoch": 26.531004366812226, + "grad_norm": 0.011034650728106499, + "learning_rate": 0.0014363937148504588, + "loss": 4.054254531860352, + "step": 1910 + }, + { + "epoch": 26.544978165938865, + "grad_norm": 0.010651151649653912, + "learning_rate": 0.0014350626825329897, + "loss": 4.037753582000732, + "step": 1911 + }, + { + "epoch": 26.558951965065503, + "grad_norm": 0.010747930034995079, + "learning_rate": 0.0014337317146195733, + "loss": 4.059938907623291, + "step": 1912 + }, + { + "epoch": 26.57292576419214, + "grad_norm": 0.00987576600164175, + "learning_rate": 0.0014324008121825468, + "loss": 4.075366973876953, + "step": 1913 + }, + { + "epoch": 26.586899563318777, + "grad_norm": 0.008504371158778667, + "learning_rate": 0.0014310699762941936, + "loss": 4.026755332946777, + "step": 1914 + }, + { + "epoch": 26.600873362445416, + "grad_norm": 0.009445052593946457, + "learning_rate": 0.001429739208026745, + "loss": 4.034431457519531, + "step": 1915 + }, + { + "epoch": 26.61484716157205, + "grad_norm": 0.010560023598372936, + "learning_rate": 0.0014284085084523763, + "loss": 4.101158618927002, + "step": 1916 + }, + { + "epoch": 26.62882096069869, + "grad_norm": 0.009920580312609673, + "learning_rate": 0.0014270778786432078, + "loss": 4.041321754455566, + "step": 1917 + }, + { + "epoch": 26.64279475982533, + "grad_norm": 0.009854268282651901, + "learning_rate": 0.0014257473196713053, + "loss": 4.049882411956787, + "step": 1918 + }, + { + "epoch": 26.656768558951963, + "grad_norm": 0.009825135581195354, + "learning_rate": 0.0014244168326086745, + "loss": 4.067440986633301, + "step": 1919 + }, + { + "epoch": 26.670742358078602, + "grad_norm": 0.010506784543395042, + "learning_rate": 0.0014230864185272663, + "loss": 4.042476654052734, + "step": 1920 + }, + { + "epoch": 26.68471615720524, + "grad_norm": 0.009527767077088356, + "learning_rate": 0.0014217560784989696, + "loss": 4.055248260498047, + "step": 1921 + }, + { + "epoch": 26.69868995633188, + "grad_norm": 0.01013826485723257, + "learning_rate": 0.001420425813595617, + "loss": 4.05246639251709, + "step": 1922 + }, + { + "epoch": 26.712663755458514, + "grad_norm": 0.011020111851394176, + "learning_rate": 0.0014190956248889778, + "loss": 4.032987117767334, + "step": 1923 + }, + { + "epoch": 26.726637554585153, + "grad_norm": 0.010647117160260677, + "learning_rate": 0.0014177655134507614, + "loss": 4.093589782714844, + "step": 1924 + }, + { + "epoch": 26.74061135371179, + "grad_norm": 0.009484780952334404, + "learning_rate": 0.001416435480352615, + "loss": 4.017437934875488, + "step": 1925 + }, + { + "epoch": 26.754585152838427, + "grad_norm": 0.0114199910312891, + "learning_rate": 0.0014151055266661211, + "loss": 4.080643653869629, + "step": 1926 + }, + { + "epoch": 26.768558951965066, + "grad_norm": 0.011137664318084717, + "learning_rate": 0.0014137756534628007, + "loss": 4.069380760192871, + "step": 1927 + }, + { + "epoch": 26.782532751091704, + "grad_norm": 0.008972682990133762, + "learning_rate": 0.0014124458618141073, + "loss": 4.051383018493652, + "step": 1928 + }, + { + "epoch": 26.79650655021834, + "grad_norm": 0.010311643593013287, + "learning_rate": 0.0014111161527914317, + "loss": 4.048953533172607, + "step": 1929 + }, + { + "epoch": 26.810480349344978, + "grad_norm": 0.01113453321158886, + "learning_rate": 0.0014097865274660944, + "loss": 4.063544273376465, + "step": 1930 + }, + { + "epoch": 26.824454148471617, + "grad_norm": 0.01091703586280346, + "learning_rate": 0.0014084569869093518, + "loss": 4.006721496582031, + "step": 1931 + }, + { + "epoch": 26.83842794759825, + "grad_norm": 0.010672449134290218, + "learning_rate": 0.0014071275321923906, + "loss": 4.020220756530762, + "step": 1932 + }, + { + "epoch": 26.85240174672489, + "grad_norm": 0.009682527743279934, + "learning_rate": 0.001405798164386328, + "loss": 4.057307243347168, + "step": 1933 + }, + { + "epoch": 26.86637554585153, + "grad_norm": 0.009808676317334175, + "learning_rate": 0.0014044688845622126, + "loss": 4.105412483215332, + "step": 1934 + }, + { + "epoch": 26.880349344978168, + "grad_norm": 0.010351189412176609, + "learning_rate": 0.0014031396937910202, + "loss": 4.066454887390137, + "step": 1935 + }, + { + "epoch": 26.894323144104803, + "grad_norm": 0.010038275271654129, + "learning_rate": 0.001401810593143656, + "loss": 4.056752681732178, + "step": 1936 + }, + { + "epoch": 26.90829694323144, + "grad_norm": 0.01016094908118248, + "learning_rate": 0.001400481583690952, + "loss": 4.097933292388916, + "step": 1937 + }, + { + "epoch": 26.92227074235808, + "grad_norm": 0.008683547377586365, + "learning_rate": 0.0013991526665036679, + "loss": 4.0386152267456055, + "step": 1938 + }, + { + "epoch": 26.936244541484715, + "grad_norm": 0.009556332603096962, + "learning_rate": 0.0013978238426524882, + "loss": 4.05835485458374, + "step": 1939 + }, + { + "epoch": 26.950218340611354, + "grad_norm": 0.009597044438123703, + "learning_rate": 0.001396495113208022, + "loss": 4.052363872528076, + "step": 1940 + }, + { + "epoch": 26.964192139737992, + "grad_norm": 0.010021139867603779, + "learning_rate": 0.0013951664792408026, + "loss": 4.083431243896484, + "step": 1941 + }, + { + "epoch": 26.978165938864628, + "grad_norm": 0.010611358098685741, + "learning_rate": 0.0013938379418212863, + "loss": 4.042330265045166, + "step": 1942 + }, + { + "epoch": 26.992139737991266, + "grad_norm": 0.009132112376391888, + "learning_rate": 0.001392509502019852, + "loss": 4.058825969696045, + "step": 1943 + }, + { + "epoch": 27.0, + "grad_norm": 0.009808709844946861, + "learning_rate": 0.0013911811609067992, + "loss": 4.093010902404785, + "step": 1944 + }, + { + "epoch": 27.0, + "eval_loss": 4.4473161697387695, + "eval_runtime": 51.7738, + "eval_samples_per_second": 47.167, + "eval_steps_per_second": 1.487, + "step": 1944 + }, + { + "epoch": 27.01397379912664, + "grad_norm": 0.011859052814543247, + "learning_rate": 0.0013898529195523488, + "loss": 3.9572651386260986, + "step": 1945 + }, + { + "epoch": 27.027947598253274, + "grad_norm": 0.013210238888859749, + "learning_rate": 0.0013885247790266406, + "loss": 3.9812350273132324, + "step": 1946 + }, + { + "epoch": 27.041921397379912, + "grad_norm": 0.012981324456632137, + "learning_rate": 0.001387196740399733, + "loss": 3.987725019454956, + "step": 1947 + }, + { + "epoch": 27.05589519650655, + "grad_norm": 0.01424382347613573, + "learning_rate": 0.001385868804741604, + "loss": 3.9719889163970947, + "step": 1948 + }, + { + "epoch": 27.069868995633186, + "grad_norm": 0.0154642965644598, + "learning_rate": 0.0013845409731221465, + "loss": 3.984248161315918, + "step": 1949 + }, + { + "epoch": 27.083842794759825, + "grad_norm": 0.014779271557927132, + "learning_rate": 0.0013832132466111706, + "loss": 4.002912521362305, + "step": 1950 + }, + { + "epoch": 27.097816593886463, + "grad_norm": 0.012921188026666641, + "learning_rate": 0.0013818856262784013, + "loss": 3.9588499069213867, + "step": 1951 + }, + { + "epoch": 27.111790393013102, + "grad_norm": 0.01165325753390789, + "learning_rate": 0.0013805581131934792, + "loss": 3.993288993835449, + "step": 1952 + }, + { + "epoch": 27.125764192139737, + "grad_norm": 0.008908954448997974, + "learning_rate": 0.0013792307084259563, + "loss": 3.987151861190796, + "step": 1953 + }, + { + "epoch": 27.139737991266376, + "grad_norm": 0.012040507979691029, + "learning_rate": 0.0013779034130452991, + "loss": 3.9675703048706055, + "step": 1954 + }, + { + "epoch": 27.153711790393015, + "grad_norm": 0.013464349322021008, + "learning_rate": 0.0013765762281208868, + "loss": 3.9901437759399414, + "step": 1955 + }, + { + "epoch": 27.16768558951965, + "grad_norm": 0.009626250714063644, + "learning_rate": 0.0013752491547220063, + "loss": 3.9373281002044678, + "step": 1956 + }, + { + "epoch": 27.18165938864629, + "grad_norm": 0.00897257961332798, + "learning_rate": 0.0013739221939178582, + "loss": 4.00804328918457, + "step": 1957 + }, + { + "epoch": 27.195633187772927, + "grad_norm": 0.010689414106309414, + "learning_rate": 0.00137259534677755, + "loss": 4.020804405212402, + "step": 1958 + }, + { + "epoch": 27.209606986899562, + "grad_norm": 0.010462156496942043, + "learning_rate": 0.0013712686143700988, + "loss": 4.028973579406738, + "step": 1959 + }, + { + "epoch": 27.2235807860262, + "grad_norm": 0.009135857224464417, + "learning_rate": 0.001369941997764429, + "loss": 3.9798216819763184, + "step": 1960 + }, + { + "epoch": 27.23755458515284, + "grad_norm": 0.009298899210989475, + "learning_rate": 0.0013686154980293713, + "loss": 4.018845081329346, + "step": 1961 + }, + { + "epoch": 27.251528384279474, + "grad_norm": 0.00979928020387888, + "learning_rate": 0.0013672891162336636, + "loss": 3.9946775436401367, + "step": 1962 + }, + { + "epoch": 27.265502183406113, + "grad_norm": 0.008175542578101158, + "learning_rate": 0.0013659628534459457, + "loss": 3.981858491897583, + "step": 1963 + }, + { + "epoch": 27.27947598253275, + "grad_norm": 0.008302642032504082, + "learning_rate": 0.0013646367107347662, + "loss": 3.991128444671631, + "step": 1964 + }, + { + "epoch": 27.29344978165939, + "grad_norm": 0.008516009896993637, + "learning_rate": 0.0013633106891685718, + "loss": 4.004990100860596, + "step": 1965 + }, + { + "epoch": 27.307423580786025, + "grad_norm": 0.007936482317745686, + "learning_rate": 0.0013619847898157164, + "loss": 3.995734214782715, + "step": 1966 + }, + { + "epoch": 27.321397379912664, + "grad_norm": 0.008317600004374981, + "learning_rate": 0.001360659013744451, + "loss": 4.040653228759766, + "step": 1967 + }, + { + "epoch": 27.335371179039303, + "grad_norm": 0.008481282740831375, + "learning_rate": 0.001359333362022931, + "loss": 3.969860553741455, + "step": 1968 + }, + { + "epoch": 27.349344978165938, + "grad_norm": 0.008944948203861713, + "learning_rate": 0.0013580078357192096, + "loss": 3.9908909797668457, + "step": 1969 + }, + { + "epoch": 27.363318777292577, + "grad_norm": 0.008897850289940834, + "learning_rate": 0.001356682435901239, + "loss": 4.04786491394043, + "step": 1970 + }, + { + "epoch": 27.377292576419215, + "grad_norm": 0.009527456946671009, + "learning_rate": 0.001355357163636871, + "loss": 4.017423629760742, + "step": 1971 + }, + { + "epoch": 27.39126637554585, + "grad_norm": 0.011714582331478596, + "learning_rate": 0.001354032019993852, + "loss": 3.9990692138671875, + "step": 1972 + }, + { + "epoch": 27.40524017467249, + "grad_norm": 0.011083807796239853, + "learning_rate": 0.0013527070060398274, + "loss": 3.995265483856201, + "step": 1973 + }, + { + "epoch": 27.419213973799128, + "grad_norm": 0.009878288954496384, + "learning_rate": 0.0013513821228423363, + "loss": 4.0364532470703125, + "step": 1974 + }, + { + "epoch": 27.433187772925763, + "grad_norm": 0.009727329947054386, + "learning_rate": 0.0013500573714688136, + "loss": 4.035953521728516, + "step": 1975 + }, + { + "epoch": 27.4471615720524, + "grad_norm": 0.009482033550739288, + "learning_rate": 0.0013487327529865879, + "loss": 4.015855312347412, + "step": 1976 + }, + { + "epoch": 27.46113537117904, + "grad_norm": 0.009037062525749207, + "learning_rate": 0.0013474082684628798, + "loss": 4.024164199829102, + "step": 1977 + }, + { + "epoch": 27.475109170305675, + "grad_norm": 0.009094092063605785, + "learning_rate": 0.0013460839189648028, + "loss": 4.037765979766846, + "step": 1978 + }, + { + "epoch": 27.489082969432314, + "grad_norm": 0.008230654522776604, + "learning_rate": 0.0013447597055593612, + "loss": 4.006317615509033, + "step": 1979 + }, + { + "epoch": 27.503056768558952, + "grad_norm": 0.008491289801895618, + "learning_rate": 0.0013434356293134501, + "loss": 4.042603969573975, + "step": 1980 + }, + { + "epoch": 27.51703056768559, + "grad_norm": 0.008907693438231945, + "learning_rate": 0.0013421116912938533, + "loss": 4.008098602294922, + "step": 1981 + }, + { + "epoch": 27.531004366812226, + "grad_norm": 0.008897818624973297, + "learning_rate": 0.0013407878925672438, + "loss": 3.9892945289611816, + "step": 1982 + }, + { + "epoch": 27.544978165938865, + "grad_norm": 0.00957555789500475, + "learning_rate": 0.0013394642342001832, + "loss": 4.046009063720703, + "step": 1983 + }, + { + "epoch": 27.558951965065503, + "grad_norm": 0.009948568418622017, + "learning_rate": 0.0013381407172591185, + "loss": 3.9760642051696777, + "step": 1984 + }, + { + "epoch": 27.57292576419214, + "grad_norm": 0.01014955248683691, + "learning_rate": 0.0013368173428103835, + "loss": 4.023215293884277, + "step": 1985 + }, + { + "epoch": 27.586899563318777, + "grad_norm": 0.008578895591199398, + "learning_rate": 0.0013354941119201966, + "loss": 4.002819061279297, + "step": 1986 + }, + { + "epoch": 27.600873362445416, + "grad_norm": 0.008729875087738037, + "learning_rate": 0.001334171025654662, + "loss": 4.0340657234191895, + "step": 1987 + }, + { + "epoch": 27.61484716157205, + "grad_norm": 0.009556280449032784, + "learning_rate": 0.0013328480850797655, + "loss": 3.983076333999634, + "step": 1988 + }, + { + "epoch": 27.62882096069869, + "grad_norm": 0.009733820334076881, + "learning_rate": 0.0013315252912613766, + "loss": 4.066578388214111, + "step": 1989 + }, + { + "epoch": 27.64279475982533, + "grad_norm": 0.009025252424180508, + "learning_rate": 0.0013302026452652474, + "loss": 3.9980697631835938, + "step": 1990 + }, + { + "epoch": 27.656768558951963, + "grad_norm": 0.008485320024192333, + "learning_rate": 0.001328880148157008, + "loss": 4.018826961517334, + "step": 1991 + }, + { + "epoch": 27.670742358078602, + "grad_norm": 0.009539696387946606, + "learning_rate": 0.0013275578010021726, + "loss": 4.011737823486328, + "step": 1992 + }, + { + "epoch": 27.68471615720524, + "grad_norm": 0.010881016030907631, + "learning_rate": 0.0013262356048661305, + "loss": 4.007291793823242, + "step": 1993 + }, + { + "epoch": 27.69868995633188, + "grad_norm": 0.010362236760556698, + "learning_rate": 0.0013249135608141523, + "loss": 3.9929394721984863, + "step": 1994 + }, + { + "epoch": 27.712663755458514, + "grad_norm": 0.010296608321368694, + "learning_rate": 0.0013235916699113849, + "loss": 4.0015130043029785, + "step": 1995 + }, + { + "epoch": 27.726637554585153, + "grad_norm": 0.011211174540221691, + "learning_rate": 0.0013222699332228517, + "loss": 4.005324363708496, + "step": 1996 + }, + { + "epoch": 27.74061135371179, + "grad_norm": 0.009743994101881981, + "learning_rate": 0.0013209483518134525, + "loss": 4.0278143882751465, + "step": 1997 + }, + { + "epoch": 27.754585152838427, + "grad_norm": 0.008563914336264133, + "learning_rate": 0.0013196269267479607, + "loss": 3.9997801780700684, + "step": 1998 + }, + { + "epoch": 27.768558951965066, + "grad_norm": 0.009078902192413807, + "learning_rate": 0.0013183056590910255, + "loss": 4.026200294494629, + "step": 1999 + }, + { + "epoch": 27.782532751091704, + "grad_norm": 0.009618895128369331, + "learning_rate": 0.0013169845499071675, + "loss": 4.027050018310547, + "step": 2000 + }, + { + "epoch": 27.79650655021834, + "grad_norm": 0.008920644409954548, + "learning_rate": 0.0013156636002607822, + "loss": 4.000490665435791, + "step": 2001 + }, + { + "epoch": 27.810480349344978, + "grad_norm": 0.008849202655255795, + "learning_rate": 0.0013143428112161323, + "loss": 4.044985294342041, + "step": 2002 + }, + { + "epoch": 27.824454148471617, + "grad_norm": 0.009303621016442776, + "learning_rate": 0.001313022183837356, + "loss": 4.046281814575195, + "step": 2003 + }, + { + "epoch": 27.83842794759825, + "grad_norm": 0.01076227892190218, + "learning_rate": 0.0013117017191884579, + "loss": 4.030576705932617, + "step": 2004 + }, + { + "epoch": 27.85240174672489, + "grad_norm": 0.010400841943919659, + "learning_rate": 0.0013103814183333127, + "loss": 3.980081558227539, + "step": 2005 + }, + { + "epoch": 27.86637554585153, + "grad_norm": 0.009459166787564754, + "learning_rate": 0.001309061282335663, + "loss": 4.014824867248535, + "step": 2006 + }, + { + "epoch": 27.880349344978168, + "grad_norm": 0.009054969064891338, + "learning_rate": 0.001307741312259118, + "loss": 4.0485944747924805, + "step": 2007 + }, + { + "epoch": 27.894323144104803, + "grad_norm": 0.009003991261124611, + "learning_rate": 0.0013064215091671545, + "loss": 3.9990320205688477, + "step": 2008 + }, + { + "epoch": 27.90829694323144, + "grad_norm": 0.008959448896348476, + "learning_rate": 0.001305101874123113, + "loss": 4.038620948791504, + "step": 2009 + }, + { + "epoch": 27.92227074235808, + "grad_norm": 0.009102243930101395, + "learning_rate": 0.0013037824081902015, + "loss": 4.030906677246094, + "step": 2010 + }, + { + "epoch": 27.936244541484715, + "grad_norm": 0.008599337190389633, + "learning_rate": 0.0013024631124314877, + "loss": 4.02648401260376, + "step": 2011 + }, + { + "epoch": 27.950218340611354, + "grad_norm": 0.008771977387368679, + "learning_rate": 0.0013011439879099056, + "loss": 4.027151107788086, + "step": 2012 + }, + { + "epoch": 27.964192139737992, + "grad_norm": 0.009082595817744732, + "learning_rate": 0.00129982503568825, + "loss": 4.0194478034973145, + "step": 2013 + }, + { + "epoch": 27.978165938864628, + "grad_norm": 0.009025354869663715, + "learning_rate": 0.0012985062568291763, + "loss": 4.044382095336914, + "step": 2014 + }, + { + "epoch": 27.992139737991266, + "grad_norm": 0.00875102635473013, + "learning_rate": 0.0012971876523952014, + "loss": 4.0257568359375, + "step": 2015 + }, + { + "epoch": 28.0, + "grad_norm": 0.009312704205513, + "learning_rate": 0.0012958692234487006, + "loss": 3.9964494705200195, + "step": 2016 + }, + { + "epoch": 28.0, + "eval_loss": 4.431707859039307, + "eval_runtime": 50.7311, + "eval_samples_per_second": 48.136, + "eval_steps_per_second": 1.518, + "step": 2016 + }, + { + "epoch": 28.01397379912664, + "grad_norm": 0.00992550328373909, + "learning_rate": 0.0012945509710519088, + "loss": 3.9253792762756348, + "step": 2017 + }, + { + "epoch": 28.027947598253274, + "grad_norm": 0.009920453652739525, + "learning_rate": 0.0012932328962669173, + "loss": 3.9438819885253906, + "step": 2018 + }, + { + "epoch": 28.041921397379912, + "grad_norm": 0.009437218308448792, + "learning_rate": 0.0012919150001556756, + "loss": 3.960963249206543, + "step": 2019 + }, + { + "epoch": 28.05589519650655, + "grad_norm": 0.009616893716156483, + "learning_rate": 0.0012905972837799896, + "loss": 3.969301223754883, + "step": 2020 + }, + { + "epoch": 28.069868995633186, + "grad_norm": 0.009107022546231747, + "learning_rate": 0.0012892797482015185, + "loss": 3.9626078605651855, + "step": 2021 + }, + { + "epoch": 28.083842794759825, + "grad_norm": 0.011317167431116104, + "learning_rate": 0.0012879623944817772, + "loss": 3.946483850479126, + "step": 2022 + }, + { + "epoch": 28.097816593886463, + "grad_norm": 0.010536770336329937, + "learning_rate": 0.001286645223682134, + "loss": 3.959566116333008, + "step": 2023 + }, + { + "epoch": 28.111790393013102, + "grad_norm": 0.009349033236503601, + "learning_rate": 0.0012853282368638097, + "loss": 3.964435338973999, + "step": 2024 + }, + { + "epoch": 28.125764192139737, + "grad_norm": 0.009950080886483192, + "learning_rate": 0.0012840114350878762, + "loss": 3.9729413986206055, + "step": 2025 + }, + { + "epoch": 28.139737991266376, + "grad_norm": 0.010833023115992546, + "learning_rate": 0.001282694819415257, + "loss": 4.006421089172363, + "step": 2026 + }, + { + "epoch": 28.153711790393015, + "grad_norm": 0.011538525111973286, + "learning_rate": 0.0012813783909067266, + "loss": 3.9558355808258057, + "step": 2027 + }, + { + "epoch": 28.16768558951965, + "grad_norm": 0.010671315714716911, + "learning_rate": 0.0012800621506229065, + "loss": 4.0052337646484375, + "step": 2028 + }, + { + "epoch": 28.18165938864629, + "grad_norm": 0.010835555382072926, + "learning_rate": 0.0012787460996242686, + "loss": 3.958641767501831, + "step": 2029 + }, + { + "epoch": 28.195633187772927, + "grad_norm": 0.01082367543131113, + "learning_rate": 0.0012774302389711304, + "loss": 3.9818902015686035, + "step": 2030 + }, + { + "epoch": 28.209606986899562, + "grad_norm": 0.009194346144795418, + "learning_rate": 0.0012761145697236583, + "loss": 3.9688854217529297, + "step": 2031 + }, + { + "epoch": 28.2235807860262, + "grad_norm": 0.007859935984015465, + "learning_rate": 0.001274799092941862, + "loss": 3.9730639457702637, + "step": 2032 + }, + { + "epoch": 28.23755458515284, + "grad_norm": 0.008534743450582027, + "learning_rate": 0.0012734838096855978, + "loss": 3.975003719329834, + "step": 2033 + }, + { + "epoch": 28.251528384279474, + "grad_norm": 0.009369445964694023, + "learning_rate": 0.0012721687210145666, + "loss": 3.972907304763794, + "step": 2034 + }, + { + "epoch": 28.265502183406113, + "grad_norm": 0.0090147340670228, + "learning_rate": 0.0012708538279883098, + "loss": 3.964256525039673, + "step": 2035 + }, + { + "epoch": 28.27947598253275, + "grad_norm": 0.007745139300823212, + "learning_rate": 0.0012695391316662151, + "loss": 3.9504833221435547, + "step": 2036 + }, + { + "epoch": 28.29344978165939, + "grad_norm": 0.008920101448893547, + "learning_rate": 0.0012682246331075075, + "loss": 3.9735097885131836, + "step": 2037 + }, + { + "epoch": 28.307423580786025, + "grad_norm": 0.008987226523458958, + "learning_rate": 0.0012669103333712568, + "loss": 3.9482927322387695, + "step": 2038 + }, + { + "epoch": 28.321397379912664, + "grad_norm": 0.00841289572417736, + "learning_rate": 0.0012655962335163685, + "loss": 3.965221881866455, + "step": 2039 + }, + { + "epoch": 28.335371179039303, + "grad_norm": 0.008016116917133331, + "learning_rate": 0.0012642823346015902, + "loss": 3.9390206336975098, + "step": 2040 + }, + { + "epoch": 28.349344978165938, + "grad_norm": 0.008470152504742146, + "learning_rate": 0.0012629686376855068, + "loss": 3.9257445335388184, + "step": 2041 + }, + { + "epoch": 28.363318777292577, + "grad_norm": 0.008637234568595886, + "learning_rate": 0.0012616551438265394, + "loss": 3.9999356269836426, + "step": 2042 + }, + { + "epoch": 28.377292576419215, + "grad_norm": 0.008036923594772816, + "learning_rate": 0.0012603418540829473, + "loss": 3.9755172729492188, + "step": 2043 + }, + { + "epoch": 28.39126637554585, + "grad_norm": 0.008908843621611595, + "learning_rate": 0.0012590287695128232, + "loss": 3.9248545169830322, + "step": 2044 + }, + { + "epoch": 28.40524017467249, + "grad_norm": 0.008743193000555038, + "learning_rate": 0.001257715891174096, + "loss": 3.9619390964508057, + "step": 2045 + }, + { + "epoch": 28.419213973799128, + "grad_norm": 0.00834199320524931, + "learning_rate": 0.001256403220124528, + "loss": 3.961569309234619, + "step": 2046 + }, + { + "epoch": 28.433187772925763, + "grad_norm": 0.009148893877863884, + "learning_rate": 0.001255090757421715, + "loss": 3.972104549407959, + "step": 2047 + }, + { + "epoch": 28.4471615720524, + "grad_norm": 0.00925084762275219, + "learning_rate": 0.0012537785041230843, + "loss": 4.000775337219238, + "step": 2048 + }, + { + "epoch": 28.46113537117904, + "grad_norm": 0.008659154176712036, + "learning_rate": 0.0012524664612858947, + "loss": 3.954564332962036, + "step": 2049 + }, + { + "epoch": 28.475109170305675, + "grad_norm": 0.007787859998643398, + "learning_rate": 0.0012511546299672355, + "loss": 3.9498860836029053, + "step": 2050 + }, + { + "epoch": 28.489082969432314, + "grad_norm": 0.00817918125540018, + "learning_rate": 0.0012498430112240246, + "loss": 4.003897666931152, + "step": 2051 + }, + { + "epoch": 28.503056768558952, + "grad_norm": 0.009282548911869526, + "learning_rate": 0.001248531606113011, + "loss": 3.9621925354003906, + "step": 2052 + }, + { + "epoch": 28.51703056768559, + "grad_norm": 0.009375610388815403, + "learning_rate": 0.001247220415690769, + "loss": 3.9977524280548096, + "step": 2053 + }, + { + "epoch": 28.531004366812226, + "grad_norm": 0.009650378488004208, + "learning_rate": 0.001245909441013701, + "loss": 3.9882829189300537, + "step": 2054 + }, + { + "epoch": 28.544978165938865, + "grad_norm": 0.009350919164717197, + "learning_rate": 0.0012445986831380366, + "loss": 4.019596099853516, + "step": 2055 + }, + { + "epoch": 28.558951965065503, + "grad_norm": 0.008726303465664387, + "learning_rate": 0.0012432881431198289, + "loss": 3.9933621883392334, + "step": 2056 + }, + { + "epoch": 28.57292576419214, + "grad_norm": 0.00829484686255455, + "learning_rate": 0.0012419778220149568, + "loss": 4.001237869262695, + "step": 2057 + }, + { + "epoch": 28.586899563318777, + "grad_norm": 0.008744452148675919, + "learning_rate": 0.0012406677208791218, + "loss": 3.9323887825012207, + "step": 2058 + }, + { + "epoch": 28.600873362445416, + "grad_norm": 0.008170492947101593, + "learning_rate": 0.0012393578407678496, + "loss": 3.925581455230713, + "step": 2059 + }, + { + "epoch": 28.61484716157205, + "grad_norm": 0.007843422703444958, + "learning_rate": 0.001238048182736486, + "loss": 4.001023769378662, + "step": 2060 + }, + { + "epoch": 28.62882096069869, + "grad_norm": 0.008914309553802013, + "learning_rate": 0.0012367387478401987, + "loss": 3.9804537296295166, + "step": 2061 + }, + { + "epoch": 28.64279475982533, + "grad_norm": 0.009736751206219196, + "learning_rate": 0.0012354295371339774, + "loss": 3.9831767082214355, + "step": 2062 + }, + { + "epoch": 28.656768558951963, + "grad_norm": 0.010061381384730339, + "learning_rate": 0.0012341205516726273, + "loss": 3.9462435245513916, + "step": 2063 + }, + { + "epoch": 28.670742358078602, + "grad_norm": 0.010350692085921764, + "learning_rate": 0.0012328117925107765, + "loss": 4.012863636016846, + "step": 2064 + }, + { + "epoch": 28.68471615720524, + "grad_norm": 0.01039902213960886, + "learning_rate": 0.0012315032607028667, + "loss": 3.9811301231384277, + "step": 2065 + }, + { + "epoch": 28.69868995633188, + "grad_norm": 0.011423371732234955, + "learning_rate": 0.0012301949573031595, + "loss": 3.9773542881011963, + "step": 2066 + }, + { + "epoch": 28.712663755458514, + "grad_norm": 0.012449614703655243, + "learning_rate": 0.001228886883365731, + "loss": 3.969600200653076, + "step": 2067 + }, + { + "epoch": 28.726637554585153, + "grad_norm": 0.010361429303884506, + "learning_rate": 0.0012275790399444726, + "loss": 4.035360336303711, + "step": 2068 + }, + { + "epoch": 28.74061135371179, + "grad_norm": 0.009603082202374935, + "learning_rate": 0.0012262714280930904, + "loss": 3.9840846061706543, + "step": 2069 + }, + { + "epoch": 28.754585152838427, + "grad_norm": 0.009183845482766628, + "learning_rate": 0.0012249640488651036, + "loss": 4.031542778015137, + "step": 2070 + }, + { + "epoch": 28.768558951965066, + "grad_norm": 0.009694446809589863, + "learning_rate": 0.0012236569033138441, + "loss": 3.9935057163238525, + "step": 2071 + }, + { + "epoch": 28.782532751091704, + "grad_norm": 0.0097816726192832, + "learning_rate": 0.0012223499924924545, + "loss": 3.9851176738739014, + "step": 2072 + }, + { + "epoch": 28.79650655021834, + "grad_norm": 0.009413184598088264, + "learning_rate": 0.001221043317453891, + "loss": 3.9642205238342285, + "step": 2073 + }, + { + "epoch": 28.810480349344978, + "grad_norm": 0.009960711002349854, + "learning_rate": 0.0012197368792509166, + "loss": 3.9373817443847656, + "step": 2074 + }, + { + "epoch": 28.824454148471617, + "grad_norm": 0.010124670341610909, + "learning_rate": 0.0012184306789361059, + "loss": 3.958115577697754, + "step": 2075 + }, + { + "epoch": 28.83842794759825, + "grad_norm": 0.011087656021118164, + "learning_rate": 0.00121712471756184, + "loss": 3.9607057571411133, + "step": 2076 + }, + { + "epoch": 28.85240174672489, + "grad_norm": 0.009812800213694572, + "learning_rate": 0.001215818996180309, + "loss": 4.00722599029541, + "step": 2077 + }, + { + "epoch": 28.86637554585153, + "grad_norm": 0.010305573232471943, + "learning_rate": 0.0012145135158435094, + "loss": 3.98728084564209, + "step": 2078 + }, + { + "epoch": 28.880349344978168, + "grad_norm": 0.01019787136465311, + "learning_rate": 0.0012132082776032422, + "loss": 3.981172800064087, + "step": 2079 + }, + { + "epoch": 28.894323144104803, + "grad_norm": 0.009260181337594986, + "learning_rate": 0.001211903282511115, + "loss": 4.034339904785156, + "step": 2080 + }, + { + "epoch": 28.90829694323144, + "grad_norm": 0.00972998607903719, + "learning_rate": 0.001210598531618538, + "loss": 4.022885322570801, + "step": 2081 + }, + { + "epoch": 28.92227074235808, + "grad_norm": 0.009853586554527283, + "learning_rate": 0.0012092940259767267, + "loss": 4.000235080718994, + "step": 2082 + }, + { + "epoch": 28.936244541484715, + "grad_norm": 0.00961449183523655, + "learning_rate": 0.0012079897666366962, + "loss": 3.9951353073120117, + "step": 2083 + }, + { + "epoch": 28.950218340611354, + "grad_norm": 0.010257176123559475, + "learning_rate": 0.0012066857546492657, + "loss": 3.9801371097564697, + "step": 2084 + }, + { + "epoch": 28.964192139737992, + "grad_norm": 0.009443561546504498, + "learning_rate": 0.001205381991065054, + "loss": 3.9596502780914307, + "step": 2085 + }, + { + "epoch": 28.978165938864628, + "grad_norm": 0.010389018803834915, + "learning_rate": 0.0012040784769344795, + "loss": 3.980527639389038, + "step": 2086 + }, + { + "epoch": 28.992139737991266, + "grad_norm": 0.010322567075490952, + "learning_rate": 0.0012027752133077607, + "loss": 3.996258497238159, + "step": 2087 + }, + { + "epoch": 29.0, + "grad_norm": 0.009987728670239449, + "learning_rate": 0.0012014722012349125, + "loss": 3.9493203163146973, + "step": 2088 + }, + { + "epoch": 29.0, + "eval_loss": 4.421086311340332, + "eval_runtime": 52.5623, + "eval_samples_per_second": 46.459, + "eval_steps_per_second": 1.465, + "step": 2088 + }, + { + "epoch": 29.01397379912664, + "grad_norm": 0.009405115619301796, + "learning_rate": 0.0012001694417657493, + "loss": 3.927506923675537, + "step": 2089 + }, + { + "epoch": 29.027947598253274, + "grad_norm": 0.009974643588066101, + "learning_rate": 0.00119886693594988, + "loss": 3.9373550415039062, + "step": 2090 + }, + { + "epoch": 29.041921397379912, + "grad_norm": 0.011331096291542053, + "learning_rate": 0.0011975646848367098, + "loss": 3.935594081878662, + "step": 2091 + }, + { + "epoch": 29.05589519650655, + "grad_norm": 0.010538448579609394, + "learning_rate": 0.0011962626894754405, + "loss": 3.911579132080078, + "step": 2092 + }, + { + "epoch": 29.069868995633186, + "grad_norm": 0.00989572424441576, + "learning_rate": 0.0011949609509150645, + "loss": 3.92525577545166, + "step": 2093 + }, + { + "epoch": 29.083842794759825, + "grad_norm": 0.010533696971833706, + "learning_rate": 0.0011936594702043704, + "loss": 3.898075580596924, + "step": 2094 + }, + { + "epoch": 29.097816593886463, + "grad_norm": 0.01051455456763506, + "learning_rate": 0.0011923582483919365, + "loss": 3.93463134765625, + "step": 2095 + }, + { + "epoch": 29.111790393013102, + "grad_norm": 0.009662075899541378, + "learning_rate": 0.0011910572865261348, + "loss": 3.898221731185913, + "step": 2096 + }, + { + "epoch": 29.125764192139737, + "grad_norm": 0.008874096907675266, + "learning_rate": 0.0011897565856551257, + "loss": 3.926787853240967, + "step": 2097 + }, + { + "epoch": 29.139737991266376, + "grad_norm": 0.008126807399094105, + "learning_rate": 0.0011884561468268606, + "loss": 3.9158413410186768, + "step": 2098 + }, + { + "epoch": 29.153711790393015, + "grad_norm": 0.007915553636848927, + "learning_rate": 0.001187155971089081, + "loss": 3.8891496658325195, + "step": 2099 + }, + { + "epoch": 29.16768558951965, + "grad_norm": 0.008312922902405262, + "learning_rate": 0.0011858560594893129, + "loss": 3.932219982147217, + "step": 2100 + }, + { + "epoch": 29.18165938864629, + "grad_norm": 0.008892328478395939, + "learning_rate": 0.0011845564130748734, + "loss": 3.8866336345672607, + "step": 2101 + }, + { + "epoch": 29.195633187772927, + "grad_norm": 0.009187526069581509, + "learning_rate": 0.0011832570328928628, + "loss": 3.9464707374572754, + "step": 2102 + }, + { + "epoch": 29.209606986899562, + "grad_norm": 0.008615494705736637, + "learning_rate": 0.0011819579199901693, + "loss": 3.9396629333496094, + "step": 2103 + }, + { + "epoch": 29.2235807860262, + "grad_norm": 0.008649308234453201, + "learning_rate": 0.0011806590754134636, + "loss": 3.9325060844421387, + "step": 2104 + }, + { + "epoch": 29.23755458515284, + "grad_norm": 0.008581082336604595, + "learning_rate": 0.0011793605002092021, + "loss": 3.9200944900512695, + "step": 2105 + }, + { + "epoch": 29.251528384279474, + "grad_norm": 0.008502699434757233, + "learning_rate": 0.0011780621954236238, + "loss": 3.9656100273132324, + "step": 2106 + }, + { + "epoch": 29.265502183406113, + "grad_norm": 0.0083576999604702, + "learning_rate": 0.0011767641621027477, + "loss": 3.9130215644836426, + "step": 2107 + }, + { + "epoch": 29.27947598253275, + "grad_norm": 0.008196269161999226, + "learning_rate": 0.0011754664012923785, + "loss": 3.942369222640991, + "step": 2108 + }, + { + "epoch": 29.29344978165939, + "grad_norm": 0.008857307955622673, + "learning_rate": 0.0011741689140380956, + "loss": 3.909585952758789, + "step": 2109 + }, + { + "epoch": 29.307423580786025, + "grad_norm": 0.008935349993407726, + "learning_rate": 0.0011728717013852635, + "loss": 3.9248366355895996, + "step": 2110 + }, + { + "epoch": 29.321397379912664, + "grad_norm": 0.008821200579404831, + "learning_rate": 0.0011715747643790211, + "loss": 3.9589288234710693, + "step": 2111 + }, + { + "epoch": 29.335371179039303, + "grad_norm": 0.00861942209303379, + "learning_rate": 0.0011702781040642885, + "loss": 3.9381728172302246, + "step": 2112 + }, + { + "epoch": 29.349344978165938, + "grad_norm": 0.009127574972808361, + "learning_rate": 0.0011689817214857613, + "loss": 3.897829532623291, + "step": 2113 + }, + { + "epoch": 29.363318777292577, + "grad_norm": 0.008008269593119621, + "learning_rate": 0.0011676856176879107, + "loss": 3.930056095123291, + "step": 2114 + }, + { + "epoch": 29.377292576419215, + "grad_norm": 0.010070208460092545, + "learning_rate": 0.0011663897937149853, + "loss": 3.892270803451538, + "step": 2115 + }, + { + "epoch": 29.39126637554585, + "grad_norm": 0.00903113093227148, + "learning_rate": 0.001165094250611006, + "loss": 3.9766440391540527, + "step": 2116 + }, + { + "epoch": 29.40524017467249, + "grad_norm": 0.008872415870428085, + "learning_rate": 0.0011637989894197692, + "loss": 3.9559712409973145, + "step": 2117 + }, + { + "epoch": 29.419213973799128, + "grad_norm": 0.009524816647171974, + "learning_rate": 0.0011625040111848427, + "loss": 3.9607882499694824, + "step": 2118 + }, + { + "epoch": 29.433187772925763, + "grad_norm": 0.009836779907345772, + "learning_rate": 0.0011612093169495678, + "loss": 3.9629592895507812, + "step": 2119 + }, + { + "epoch": 29.4471615720524, + "grad_norm": 0.009586540050804615, + "learning_rate": 0.0011599149077570561, + "loss": 3.95208740234375, + "step": 2120 + }, + { + "epoch": 29.46113537117904, + "grad_norm": 0.010495901107788086, + "learning_rate": 0.0011586207846501896, + "loss": 3.9439730644226074, + "step": 2121 + }, + { + "epoch": 29.475109170305675, + "grad_norm": 0.010346266441047192, + "learning_rate": 0.0011573269486716202, + "loss": 3.959299087524414, + "step": 2122 + }, + { + "epoch": 29.489082969432314, + "grad_norm": 0.010767634026706219, + "learning_rate": 0.0011560334008637674, + "loss": 3.900784969329834, + "step": 2123 + }, + { + "epoch": 29.503056768558952, + "grad_norm": 0.012122200801968575, + "learning_rate": 0.0011547401422688202, + "loss": 3.9588193893432617, + "step": 2124 + }, + { + "epoch": 29.51703056768559, + "grad_norm": 0.012914645485579967, + "learning_rate": 0.0011534471739287331, + "loss": 3.9278407096862793, + "step": 2125 + }, + { + "epoch": 29.531004366812226, + "grad_norm": 0.011785333044826984, + "learning_rate": 0.0011521544968852273, + "loss": 3.992793560028076, + "step": 2126 + }, + { + "epoch": 29.544978165938865, + "grad_norm": 0.008455781266093254, + "learning_rate": 0.0011508621121797899, + "loss": 3.96525239944458, + "step": 2127 + }, + { + "epoch": 29.558951965065503, + "grad_norm": 0.009655282832682133, + "learning_rate": 0.0011495700208536714, + "loss": 3.9346890449523926, + "step": 2128 + }, + { + "epoch": 29.57292576419214, + "grad_norm": 0.010537004098296165, + "learning_rate": 0.0011482782239478871, + "loss": 3.9534668922424316, + "step": 2129 + }, + { + "epoch": 29.586899563318777, + "grad_norm": 0.009438651613891125, + "learning_rate": 0.0011469867225032135, + "loss": 3.983487606048584, + "step": 2130 + }, + { + "epoch": 29.600873362445416, + "grad_norm": 0.010586646385490894, + "learning_rate": 0.0011456955175601908, + "loss": 3.939680576324463, + "step": 2131 + }, + { + "epoch": 29.61484716157205, + "grad_norm": 0.009082718752324581, + "learning_rate": 0.001144404610159119, + "loss": 3.9558515548706055, + "step": 2132 + }, + { + "epoch": 29.62882096069869, + "grad_norm": 0.009833602234721184, + "learning_rate": 0.0011431140013400591, + "loss": 3.9539828300476074, + "step": 2133 + }, + { + "epoch": 29.64279475982533, + "grad_norm": 0.009142170660197735, + "learning_rate": 0.0011418236921428313, + "loss": 3.9442849159240723, + "step": 2134 + }, + { + "epoch": 29.656768558951963, + "grad_norm": 0.008603853173553944, + "learning_rate": 0.0011405336836070137, + "loss": 3.9526913166046143, + "step": 2135 + }, + { + "epoch": 29.670742358078602, + "grad_norm": 0.008838302455842495, + "learning_rate": 0.0011392439767719446, + "loss": 3.973104953765869, + "step": 2136 + }, + { + "epoch": 29.68471615720524, + "grad_norm": 0.008867006748914719, + "learning_rate": 0.0011379545726767156, + "loss": 3.9832305908203125, + "step": 2137 + }, + { + "epoch": 29.69868995633188, + "grad_norm": 0.00971697736531496, + "learning_rate": 0.001136665472360178, + "loss": 3.9563937187194824, + "step": 2138 + }, + { + "epoch": 29.712663755458514, + "grad_norm": 0.00957175437361002, + "learning_rate": 0.001135376676860935, + "loss": 3.975194215774536, + "step": 2139 + }, + { + "epoch": 29.726637554585153, + "grad_norm": 0.008548660203814507, + "learning_rate": 0.001134088187217347, + "loss": 3.9839985370635986, + "step": 2140 + }, + { + "epoch": 29.74061135371179, + "grad_norm": 0.008556334301829338, + "learning_rate": 0.0011328000044675264, + "loss": 3.9225516319274902, + "step": 2141 + }, + { + "epoch": 29.754585152838427, + "grad_norm": 0.008440867997705936, + "learning_rate": 0.0011315121296493385, + "loss": 3.945028066635132, + "step": 2142 + }, + { + "epoch": 29.768558951965066, + "grad_norm": 0.009643333032727242, + "learning_rate": 0.0011302245638004017, + "loss": 3.943120002746582, + "step": 2143 + }, + { + "epoch": 29.782532751091704, + "grad_norm": 0.008832835592329502, + "learning_rate": 0.001128937307958083, + "loss": 3.9895920753479004, + "step": 2144 + }, + { + "epoch": 29.79650655021834, + "grad_norm": 0.008720474317669868, + "learning_rate": 0.0011276503631595036, + "loss": 3.941950798034668, + "step": 2145 + }, + { + "epoch": 29.810480349344978, + "grad_norm": 0.008099926635622978, + "learning_rate": 0.001126363730441529, + "loss": 3.9626870155334473, + "step": 2146 + }, + { + "epoch": 29.824454148471617, + "grad_norm": 0.008525107055902481, + "learning_rate": 0.0011250774108407778, + "loss": 3.950535774230957, + "step": 2147 + }, + { + "epoch": 29.83842794759825, + "grad_norm": 0.008998272009193897, + "learning_rate": 0.0011237914053936135, + "loss": 3.9460678100585938, + "step": 2148 + }, + { + "epoch": 29.85240174672489, + "grad_norm": 0.009387437254190445, + "learning_rate": 0.001122505715136148, + "loss": 3.926604986190796, + "step": 2149 + }, + { + "epoch": 29.86637554585153, + "grad_norm": 0.007804017513990402, + "learning_rate": 0.0011212203411042386, + "loss": 3.9894230365753174, + "step": 2150 + }, + { + "epoch": 29.880349344978168, + "grad_norm": 0.008564616553485394, + "learning_rate": 0.0011199352843334876, + "loss": 3.9339959621429443, + "step": 2151 + }, + { + "epoch": 29.894323144104803, + "grad_norm": 0.008726440370082855, + "learning_rate": 0.0011186505458592427, + "loss": 3.9867987632751465, + "step": 2152 + }, + { + "epoch": 29.90829694323144, + "grad_norm": 0.008074100129306316, + "learning_rate": 0.0011173661267165932, + "loss": 3.9492292404174805, + "step": 2153 + }, + { + "epoch": 29.92227074235808, + "grad_norm": 0.007688041776418686, + "learning_rate": 0.0011160820279403744, + "loss": 3.9425454139709473, + "step": 2154 + }, + { + "epoch": 29.936244541484715, + "grad_norm": 0.007808993570506573, + "learning_rate": 0.0011147982505651595, + "loss": 3.935114860534668, + "step": 2155 + }, + { + "epoch": 29.950218340611354, + "grad_norm": 0.009786820039153099, + "learning_rate": 0.0011135147956252659, + "loss": 3.959758758544922, + "step": 2156 + }, + { + "epoch": 29.964192139737992, + "grad_norm": 0.009487155824899673, + "learning_rate": 0.00111223166415475, + "loss": 3.989048480987549, + "step": 2157 + }, + { + "epoch": 29.978165938864628, + "grad_norm": 0.008407804183661938, + "learning_rate": 0.0011109488571874073, + "loss": 3.932347297668457, + "step": 2158 + }, + { + "epoch": 29.992139737991266, + "grad_norm": 0.007158380467444658, + "learning_rate": 0.0011096663757567723, + "loss": 3.9545154571533203, + "step": 2159 + }, + { + "epoch": 30.0, + "grad_norm": 0.008840581402182579, + "learning_rate": 0.001108384220896117, + "loss": 3.965984344482422, + "step": 2160 + }, + { + "epoch": 30.0, + "eval_loss": 4.409194469451904, + "eval_runtime": 51.5561, + "eval_samples_per_second": 47.366, + "eval_steps_per_second": 1.494, + "step": 2160 + }, + { + "epoch": 30.01397379912664, + "grad_norm": 0.010702896863222122, + "learning_rate": 0.0011071023936384508, + "loss": 3.8964462280273438, + "step": 2161 + }, + { + "epoch": 30.027947598253274, + "grad_norm": 0.01290619745850563, + "learning_rate": 0.0011058208950165183, + "loss": 3.8775625228881836, + "step": 2162 + }, + { + "epoch": 30.041921397379912, + "grad_norm": 0.01443667896091938, + "learning_rate": 0.0011045397260628, + "loss": 3.8755288124084473, + "step": 2163 + }, + { + "epoch": 30.05589519650655, + "grad_norm": 0.014411968179047108, + "learning_rate": 0.0011032588878095108, + "loss": 3.889918327331543, + "step": 2164 + }, + { + "epoch": 30.069868995633186, + "grad_norm": 0.013656266033649445, + "learning_rate": 0.0011019783812885986, + "loss": 3.926997661590576, + "step": 2165 + }, + { + "epoch": 30.083842794759825, + "grad_norm": 0.010661937296390533, + "learning_rate": 0.0011006982075317446, + "loss": 3.9226956367492676, + "step": 2166 + }, + { + "epoch": 30.097816593886463, + "grad_norm": 0.008968768641352654, + "learning_rate": 0.0010994183675703615, + "loss": 3.8932647705078125, + "step": 2167 + }, + { + "epoch": 30.111790393013102, + "grad_norm": 0.009992181323468685, + "learning_rate": 0.0010981388624355937, + "loss": 3.8920068740844727, + "step": 2168 + }, + { + "epoch": 30.125764192139737, + "grad_norm": 0.011034460738301277, + "learning_rate": 0.0010968596931583148, + "loss": 3.909073829650879, + "step": 2169 + }, + { + "epoch": 30.139737991266376, + "grad_norm": 0.011815821751952171, + "learning_rate": 0.0010955808607691284, + "loss": 3.9190006256103516, + "step": 2170 + }, + { + "epoch": 30.153711790393015, + "grad_norm": 0.012317490763962269, + "learning_rate": 0.0010943023662983674, + "loss": 3.8790552616119385, + "step": 2171 + }, + { + "epoch": 30.16768558951965, + "grad_norm": 0.010983958840370178, + "learning_rate": 0.0010930242107760903, + "loss": 3.8730852603912354, + "step": 2172 + }, + { + "epoch": 30.18165938864629, + "grad_norm": 0.008948802947998047, + "learning_rate": 0.0010917463952320856, + "loss": 3.923659563064575, + "step": 2173 + }, + { + "epoch": 30.195633187772927, + "grad_norm": 0.009568693116307259, + "learning_rate": 0.0010904689206958649, + "loss": 3.9100794792175293, + "step": 2174 + }, + { + "epoch": 30.209606986899562, + "grad_norm": 0.011862963438034058, + "learning_rate": 0.0010891917881966672, + "loss": 3.908093214035034, + "step": 2175 + }, + { + "epoch": 30.2235807860262, + "grad_norm": 0.010180995799601078, + "learning_rate": 0.0010879149987634545, + "loss": 3.900660514831543, + "step": 2176 + }, + { + "epoch": 30.23755458515284, + "grad_norm": 0.009372464381158352, + "learning_rate": 0.0010866385534249131, + "loss": 3.8950703144073486, + "step": 2177 + }, + { + "epoch": 30.251528384279474, + "grad_norm": 0.010695629753172398, + "learning_rate": 0.0010853624532094525, + "loss": 3.913530111312866, + "step": 2178 + }, + { + "epoch": 30.265502183406113, + "grad_norm": 0.011817459017038345, + "learning_rate": 0.0010840866991452024, + "loss": 3.90268611907959, + "step": 2179 + }, + { + "epoch": 30.27947598253275, + "grad_norm": 0.010686337016522884, + "learning_rate": 0.0010828112922600167, + "loss": 3.9141364097595215, + "step": 2180 + }, + { + "epoch": 30.29344978165939, + "grad_norm": 0.009550466202199459, + "learning_rate": 0.0010815362335814655, + "loss": 3.968468189239502, + "step": 2181 + }, + { + "epoch": 30.307423580786025, + "grad_norm": 0.009367208927869797, + "learning_rate": 0.0010802615241368423, + "loss": 3.8933968544006348, + "step": 2182 + }, + { + "epoch": 30.321397379912664, + "grad_norm": 0.009899397380650043, + "learning_rate": 0.0010789871649531562, + "loss": 3.936708450317383, + "step": 2183 + }, + { + "epoch": 30.335371179039303, + "grad_norm": 0.009181005880236626, + "learning_rate": 0.0010777131570571363, + "loss": 3.9370503425598145, + "step": 2184 + }, + { + "epoch": 30.349344978165938, + "grad_norm": 0.00799303874373436, + "learning_rate": 0.0010764395014752274, + "loss": 3.911050319671631, + "step": 2185 + }, + { + "epoch": 30.363318777292577, + "grad_norm": 0.008469315245747566, + "learning_rate": 0.0010751661992335908, + "loss": 3.9103240966796875, + "step": 2186 + }, + { + "epoch": 30.377292576419215, + "grad_norm": 0.009814459830522537, + "learning_rate": 0.0010738932513581034, + "loss": 3.881843090057373, + "step": 2187 + }, + { + "epoch": 30.39126637554585, + "grad_norm": 0.010244108736515045, + "learning_rate": 0.0010726206588743557, + "loss": 3.883141040802002, + "step": 2188 + }, + { + "epoch": 30.40524017467249, + "grad_norm": 0.008898732252418995, + "learning_rate": 0.001071348422807653, + "loss": 3.952347755432129, + "step": 2189 + }, + { + "epoch": 30.419213973799128, + "grad_norm": 0.00823870487511158, + "learning_rate": 0.0010700765441830121, + "loss": 3.932507038116455, + "step": 2190 + }, + { + "epoch": 30.433187772925763, + "grad_norm": 0.009857657365500927, + "learning_rate": 0.001068805024025164, + "loss": 3.932932138442993, + "step": 2191 + }, + { + "epoch": 30.4471615720524, + "grad_norm": 0.00929308496415615, + "learning_rate": 0.0010675338633585488, + "loss": 3.8971338272094727, + "step": 2192 + }, + { + "epoch": 30.46113537117904, + "grad_norm": 0.008667961694300175, + "learning_rate": 0.0010662630632073172, + "loss": 3.916931629180908, + "step": 2193 + }, + { + "epoch": 30.475109170305675, + "grad_norm": 0.009758287109434605, + "learning_rate": 0.001064992624595331, + "loss": 3.964726209640503, + "step": 2194 + }, + { + "epoch": 30.489082969432314, + "grad_norm": 0.009429254569113255, + "learning_rate": 0.0010637225485461586, + "loss": 3.917572498321533, + "step": 2195 + }, + { + "epoch": 30.503056768558952, + "grad_norm": 0.008303353562951088, + "learning_rate": 0.0010624528360830782, + "loss": 3.875241756439209, + "step": 2196 + }, + { + "epoch": 30.51703056768559, + "grad_norm": 0.008729013614356518, + "learning_rate": 0.001061183488229073, + "loss": 3.9691600799560547, + "step": 2197 + }, + { + "epoch": 30.531004366812226, + "grad_norm": 0.010239302180707455, + "learning_rate": 0.0010599145060068353, + "loss": 3.9053149223327637, + "step": 2198 + }, + { + "epoch": 30.544978165938865, + "grad_norm": 0.00932721421122551, + "learning_rate": 0.0010586458904387596, + "loss": 3.8593955039978027, + "step": 2199 + }, + { + "epoch": 30.558951965065503, + "grad_norm": 0.008196988143026829, + "learning_rate": 0.0010573776425469473, + "loss": 3.917983055114746, + "step": 2200 + }, + { + "epoch": 30.57292576419214, + "grad_norm": 0.008049141615629196, + "learning_rate": 0.001056109763353203, + "loss": 3.9337785243988037, + "step": 2201 + }, + { + "epoch": 30.586899563318777, + "grad_norm": 0.009256313554942608, + "learning_rate": 0.001054842253879034, + "loss": 3.937887668609619, + "step": 2202 + }, + { + "epoch": 30.600873362445416, + "grad_norm": 0.009292179718613625, + "learning_rate": 0.0010535751151456495, + "loss": 3.9203450679779053, + "step": 2203 + }, + { + "epoch": 30.61484716157205, + "grad_norm": 0.009142161346971989, + "learning_rate": 0.0010523083481739602, + "loss": 3.9027457237243652, + "step": 2204 + }, + { + "epoch": 30.62882096069869, + "grad_norm": 0.009498288854956627, + "learning_rate": 0.0010510419539845777, + "loss": 3.917316436767578, + "step": 2205 + }, + { + "epoch": 30.64279475982533, + "grad_norm": 0.00849096104502678, + "learning_rate": 0.001049775933597813, + "loss": 3.9229235649108887, + "step": 2206 + }, + { + "epoch": 30.656768558951963, + "grad_norm": 0.008343606255948544, + "learning_rate": 0.0010485102880336747, + "loss": 3.9138073921203613, + "step": 2207 + }, + { + "epoch": 30.670742358078602, + "grad_norm": 0.010342134162783623, + "learning_rate": 0.0010472450183118727, + "loss": 3.8932929039001465, + "step": 2208 + }, + { + "epoch": 30.68471615720524, + "grad_norm": 0.010298306122422218, + "learning_rate": 0.00104598012545181, + "loss": 3.9112677574157715, + "step": 2209 + }, + { + "epoch": 30.69868995633188, + "grad_norm": 0.00827560480684042, + "learning_rate": 0.001044715610472589, + "loss": 3.894556760787964, + "step": 2210 + }, + { + "epoch": 30.712663755458514, + "grad_norm": 0.008993102237582207, + "learning_rate": 0.0010434514743930064, + "loss": 3.905391216278076, + "step": 2211 + }, + { + "epoch": 30.726637554585153, + "grad_norm": 0.008632538840174675, + "learning_rate": 0.0010421877182315536, + "loss": 3.896714210510254, + "step": 2212 + }, + { + "epoch": 30.74061135371179, + "grad_norm": 0.008542424067854881, + "learning_rate": 0.0010409243430064162, + "loss": 3.9145569801330566, + "step": 2213 + }, + { + "epoch": 30.754585152838427, + "grad_norm": 0.008970450609922409, + "learning_rate": 0.0010396613497354725, + "loss": 3.9170360565185547, + "step": 2214 + }, + { + "epoch": 30.768558951965066, + "grad_norm": 0.008744142949581146, + "learning_rate": 0.0010383987394362942, + "loss": 3.927412509918213, + "step": 2215 + }, + { + "epoch": 30.782532751091704, + "grad_norm": 0.008229847066104412, + "learning_rate": 0.0010371365131261424, + "loss": 3.928352117538452, + "step": 2216 + }, + { + "epoch": 30.79650655021834, + "grad_norm": 0.00818693358451128, + "learning_rate": 0.0010358746718219717, + "loss": 3.9219343662261963, + "step": 2217 + }, + { + "epoch": 30.810480349344978, + "grad_norm": 0.008848588913679123, + "learning_rate": 0.001034613216540423, + "loss": 3.9346699714660645, + "step": 2218 + }, + { + "epoch": 30.824454148471617, + "grad_norm": 0.00834780652076006, + "learning_rate": 0.0010333521482978295, + "loss": 3.9144492149353027, + "step": 2219 + }, + { + "epoch": 30.83842794759825, + "grad_norm": 0.00940527394413948, + "learning_rate": 0.0010320914681102098, + "loss": 3.951573371887207, + "step": 2220 + }, + { + "epoch": 30.85240174672489, + "grad_norm": 0.008884521201252937, + "learning_rate": 0.0010308311769932725, + "loss": 3.899625778198242, + "step": 2221 + }, + { + "epoch": 30.86637554585153, + "grad_norm": 0.007863261736929417, + "learning_rate": 0.0010295712759624107, + "loss": 3.9321727752685547, + "step": 2222 + }, + { + "epoch": 30.880349344978168, + "grad_norm": 0.008651153184473515, + "learning_rate": 0.0010283117660327035, + "loss": 3.9030094146728516, + "step": 2223 + }, + { + "epoch": 30.894323144104803, + "grad_norm": 0.00822074618190527, + "learning_rate": 0.0010270526482189163, + "loss": 3.939466953277588, + "step": 2224 + }, + { + "epoch": 30.90829694323144, + "grad_norm": 0.008701153099536896, + "learning_rate": 0.0010257939235354964, + "loss": 3.9099626541137695, + "step": 2225 + }, + { + "epoch": 30.92227074235808, + "grad_norm": 0.008418620564043522, + "learning_rate": 0.0010245355929965774, + "loss": 3.914116382598877, + "step": 2226 + }, + { + "epoch": 30.936244541484715, + "grad_norm": 0.008339700289070606, + "learning_rate": 0.001023277657615971, + "loss": 3.9111878871917725, + "step": 2227 + }, + { + "epoch": 30.950218340611354, + "grad_norm": 0.008304647170007229, + "learning_rate": 0.0010220201184071747, + "loss": 3.950974464416504, + "step": 2228 + }, + { + "epoch": 30.964192139737992, + "grad_norm": 0.008436229079961777, + "learning_rate": 0.001020762976383365, + "loss": 3.9631457328796387, + "step": 2229 + }, + { + "epoch": 30.978165938864628, + "grad_norm": 0.0074487063102424145, + "learning_rate": 0.001019506232557398, + "loss": 3.902984619140625, + "step": 2230 + }, + { + "epoch": 30.992139737991266, + "grad_norm": 0.008736515417695045, + "learning_rate": 0.0010182498879418096, + "loss": 3.9268949031829834, + "step": 2231 + }, + { + "epoch": 31.0, + "grad_norm": 0.009666468016803265, + "learning_rate": 0.0010169939435488145, + "loss": 3.9274301528930664, + "step": 2232 + }, + { + "epoch": 31.0, + "eval_loss": 4.395970821380615, + "eval_runtime": 53.7848, + "eval_samples_per_second": 45.403, + "eval_steps_per_second": 1.432, + "step": 2232 + }, + { + "epoch": 31.01397379912664, + "grad_norm": 0.009094545617699623, + "learning_rate": 0.001015738400390304, + "loss": 3.866425037384033, + "step": 2233 + }, + { + "epoch": 31.027947598253274, + "grad_norm": 0.009273733012378216, + "learning_rate": 0.0010144832594778462, + "loss": 3.8569841384887695, + "step": 2234 + }, + { + "epoch": 31.041921397379912, + "grad_norm": 0.009205271489918232, + "learning_rate": 0.0010132285218226855, + "loss": 3.8981378078460693, + "step": 2235 + }, + { + "epoch": 31.05589519650655, + "grad_norm": 0.009963045828044415, + "learning_rate": 0.0010119741884357424, + "loss": 3.881542444229126, + "step": 2236 + }, + { + "epoch": 31.069868995633186, + "grad_norm": 0.008536554872989655, + "learning_rate": 0.0010107202603276096, + "loss": 3.851253032684326, + "step": 2237 + }, + { + "epoch": 31.083842794759825, + "grad_norm": 0.01074561569839716, + "learning_rate": 0.001009466738508555, + "loss": 3.8731184005737305, + "step": 2238 + }, + { + "epoch": 31.097816593886463, + "grad_norm": 0.010809902101755142, + "learning_rate": 0.0010082136239885177, + "loss": 3.863889217376709, + "step": 2239 + }, + { + "epoch": 31.111790393013102, + "grad_norm": 0.008289352990686893, + "learning_rate": 0.0010069609177771103, + "loss": 3.839066743850708, + "step": 2240 + }, + { + "epoch": 31.125764192139737, + "grad_norm": 0.00792208593338728, + "learning_rate": 0.001005708620883615, + "loss": 3.8634843826293945, + "step": 2241 + }, + { + "epoch": 31.139737991266376, + "grad_norm": 0.008381598629057407, + "learning_rate": 0.0010044567343169845, + "loss": 3.868283987045288, + "step": 2242 + }, + { + "epoch": 31.153711790393015, + "grad_norm": 0.009784617461264133, + "learning_rate": 0.0010032052590858425, + "loss": 3.853074550628662, + "step": 2243 + }, + { + "epoch": 31.16768558951965, + "grad_norm": 0.009462198242545128, + "learning_rate": 0.0010019541961984782, + "loss": 3.889979600906372, + "step": 2244 + }, + { + "epoch": 31.18165938864629, + "grad_norm": 0.010207819752395153, + "learning_rate": 0.001000703546662852, + "loss": 3.8636741638183594, + "step": 2245 + }, + { + "epoch": 31.195633187772927, + "grad_norm": 0.009123551659286022, + "learning_rate": 0.0009994533114865887, + "loss": 3.8578555583953857, + "step": 2246 + }, + { + "epoch": 31.209606986899562, + "grad_norm": 0.009006140753626823, + "learning_rate": 0.0009982034916769806, + "loss": 3.8828015327453613, + "step": 2247 + }, + { + "epoch": 31.2235807860262, + "grad_norm": 0.0091114342212677, + "learning_rate": 0.0009969540882409846, + "loss": 3.880486011505127, + "step": 2248 + }, + { + "epoch": 31.23755458515284, + "grad_norm": 0.008851367980241776, + "learning_rate": 0.0009957051021852227, + "loss": 3.8618125915527344, + "step": 2249 + }, + { + "epoch": 31.251528384279474, + "grad_norm": 0.008196741342544556, + "learning_rate": 0.0009944565345159806, + "loss": 3.8314175605773926, + "step": 2250 + }, + { + "epoch": 31.265502183406113, + "grad_norm": 0.00863513071089983, + "learning_rate": 0.000993208386239206, + "loss": 3.856212615966797, + "step": 2251 + }, + { + "epoch": 31.27947598253275, + "grad_norm": 0.008624681271612644, + "learning_rate": 0.0009919606583605107, + "loss": 3.868927001953125, + "step": 2252 + }, + { + "epoch": 31.29344978165939, + "grad_norm": 0.009208842180669308, + "learning_rate": 0.0009907133518851647, + "loss": 3.9072189331054688, + "step": 2253 + }, + { + "epoch": 31.307423580786025, + "grad_norm": 0.010200913064181805, + "learning_rate": 0.0009894664678181027, + "loss": 3.875272750854492, + "step": 2254 + }, + { + "epoch": 31.321397379912664, + "grad_norm": 0.009892710484564304, + "learning_rate": 0.000988220007163914, + "loss": 3.888686418533325, + "step": 2255 + }, + { + "epoch": 31.335371179039303, + "grad_norm": 0.009873585775494576, + "learning_rate": 0.000986973970926851, + "loss": 3.8607118129730225, + "step": 2256 + }, + { + "epoch": 31.349344978165938, + "grad_norm": 0.009322221390902996, + "learning_rate": 0.0009857283601108228, + "loss": 3.8852052688598633, + "step": 2257 + }, + { + "epoch": 31.363318777292577, + "grad_norm": 0.008141528815031052, + "learning_rate": 0.0009844831757193947, + "loss": 3.888213872909546, + "step": 2258 + }, + { + "epoch": 31.377292576419215, + "grad_norm": 0.008191613480448723, + "learning_rate": 0.00098323841875579, + "loss": 3.9073967933654785, + "step": 2259 + }, + { + "epoch": 31.39126637554585, + "grad_norm": 0.008419081568717957, + "learning_rate": 0.0009819940902228864, + "loss": 3.879223346710205, + "step": 2260 + }, + { + "epoch": 31.40524017467249, + "grad_norm": 0.009037869982421398, + "learning_rate": 0.0009807501911232171, + "loss": 3.883303165435791, + "step": 2261 + }, + { + "epoch": 31.419213973799128, + "grad_norm": 0.00854078121483326, + "learning_rate": 0.0009795067224589687, + "loss": 3.926767587661743, + "step": 2262 + }, + { + "epoch": 31.433187772925763, + "grad_norm": 0.008590771816670895, + "learning_rate": 0.0009782636852319825, + "loss": 3.913999557495117, + "step": 2263 + }, + { + "epoch": 31.4471615720524, + "grad_norm": 0.00859815627336502, + "learning_rate": 0.0009770210804437498, + "loss": 3.8828063011169434, + "step": 2264 + }, + { + "epoch": 31.46113537117904, + "grad_norm": 0.008970260620117188, + "learning_rate": 0.0009757789090954157, + "loss": 3.9174866676330566, + "step": 2265 + }, + { + "epoch": 31.475109170305675, + "grad_norm": 0.00877175759524107, + "learning_rate": 0.0009745371721877754, + "loss": 3.8879401683807373, + "step": 2266 + }, + { + "epoch": 31.489082969432314, + "grad_norm": 0.008541764691472054, + "learning_rate": 0.0009732958707212731, + "loss": 3.8717427253723145, + "step": 2267 + }, + { + "epoch": 31.503056768558952, + "grad_norm": 0.008466241881251335, + "learning_rate": 0.0009720550056960039, + "loss": 3.8944485187530518, + "step": 2268 + }, + { + "epoch": 31.51703056768559, + "grad_norm": 0.00808491837233305, + "learning_rate": 0.0009708145781117096, + "loss": 3.8617026805877686, + "step": 2269 + }, + { + "epoch": 31.531004366812226, + "grad_norm": 0.00842207670211792, + "learning_rate": 0.0009695745889677805, + "loss": 3.887824535369873, + "step": 2270 + }, + { + "epoch": 31.544978165938865, + "grad_norm": 0.00830599945038557, + "learning_rate": 0.0009683350392632533, + "loss": 3.8612070083618164, + "step": 2271 + }, + { + "epoch": 31.558951965065503, + "grad_norm": 0.008417914621531963, + "learning_rate": 0.0009670959299968111, + "loss": 3.880140781402588, + "step": 2272 + }, + { + "epoch": 31.57292576419214, + "grad_norm": 0.008542059920728207, + "learning_rate": 0.0009658572621667822, + "loss": 3.8951029777526855, + "step": 2273 + }, + { + "epoch": 31.586899563318777, + "grad_norm": 0.008822236210107803, + "learning_rate": 0.000964619036771138, + "loss": 3.8962583541870117, + "step": 2274 + }, + { + "epoch": 31.600873362445416, + "grad_norm": 0.009151184931397438, + "learning_rate": 0.0009633812548074952, + "loss": 3.882500648498535, + "step": 2275 + }, + { + "epoch": 31.61484716157205, + "grad_norm": 0.007803036365658045, + "learning_rate": 0.0009621439172731115, + "loss": 3.8740782737731934, + "step": 2276 + }, + { + "epoch": 31.62882096069869, + "grad_norm": 0.009350850246846676, + "learning_rate": 0.0009609070251648884, + "loss": 3.8627419471740723, + "step": 2277 + }, + { + "epoch": 31.64279475982533, + "grad_norm": 0.008319144137203693, + "learning_rate": 0.0009596705794793667, + "loss": 3.9084486961364746, + "step": 2278 + }, + { + "epoch": 31.656768558951963, + "grad_norm": 0.009198528714478016, + "learning_rate": 0.0009584345812127282, + "loss": 3.92141056060791, + "step": 2279 + }, + { + "epoch": 31.670742358078602, + "grad_norm": 0.010480938479304314, + "learning_rate": 0.0009571990313607958, + "loss": 3.8916103839874268, + "step": 2280 + }, + { + "epoch": 31.68471615720524, + "grad_norm": 0.01045288611203432, + "learning_rate": 0.0009559639309190278, + "loss": 3.8863210678100586, + "step": 2281 + }, + { + "epoch": 31.69868995633188, + "grad_norm": 0.009121391922235489, + "learning_rate": 0.0009547292808825239, + "loss": 3.9167044162750244, + "step": 2282 + }, + { + "epoch": 31.712663755458514, + "grad_norm": 0.01093223039060831, + "learning_rate": 0.0009534950822460184, + "loss": 3.9046711921691895, + "step": 2283 + }, + { + "epoch": 31.726637554585153, + "grad_norm": 0.01214868575334549, + "learning_rate": 0.0009522613360038835, + "loss": 3.9126768112182617, + "step": 2284 + }, + { + "epoch": 31.74061135371179, + "grad_norm": 0.01055280864238739, + "learning_rate": 0.0009510280431501255, + "loss": 3.9214353561401367, + "step": 2285 + }, + { + "epoch": 31.754585152838427, + "grad_norm": 0.009021313861012459, + "learning_rate": 0.0009497952046783867, + "loss": 3.8795580863952637, + "step": 2286 + }, + { + "epoch": 31.768558951965066, + "grad_norm": 0.00863752793520689, + "learning_rate": 0.0009485628215819428, + "loss": 3.893232822418213, + "step": 2287 + }, + { + "epoch": 31.782532751091704, + "grad_norm": 0.009327626787126064, + "learning_rate": 0.0009473308948537021, + "loss": 3.917377471923828, + "step": 2288 + }, + { + "epoch": 31.79650655021834, + "grad_norm": 0.01003898587077856, + "learning_rate": 0.0009460994254862066, + "loss": 3.8818931579589844, + "step": 2289 + }, + { + "epoch": 31.810480349344978, + "grad_norm": 0.00892532430589199, + "learning_rate": 0.0009448684144716278, + "loss": 3.8761978149414062, + "step": 2290 + }, + { + "epoch": 31.824454148471617, + "grad_norm": 0.008396463468670845, + "learning_rate": 0.0009436378628017702, + "loss": 3.8665928840637207, + "step": 2291 + }, + { + "epoch": 31.83842794759825, + "grad_norm": 0.00784370768815279, + "learning_rate": 0.0009424077714680662, + "loss": 3.8850138187408447, + "step": 2292 + }, + { + "epoch": 31.85240174672489, + "grad_norm": 0.00809321366250515, + "learning_rate": 0.0009411781414615784, + "loss": 3.917078733444214, + "step": 2293 + }, + { + "epoch": 31.86637554585153, + "grad_norm": 0.008685288950800896, + "learning_rate": 0.000939948973772998, + "loss": 3.8865013122558594, + "step": 2294 + }, + { + "epoch": 31.880349344978168, + "grad_norm": 0.007291820365935564, + "learning_rate": 0.0009387202693926425, + "loss": 3.8967666625976562, + "step": 2295 + }, + { + "epoch": 31.894323144104803, + "grad_norm": 0.008459319360554218, + "learning_rate": 0.0009374920293104573, + "loss": 3.9043235778808594, + "step": 2296 + }, + { + "epoch": 31.90829694323144, + "grad_norm": 0.007924137637019157, + "learning_rate": 0.0009362642545160123, + "loss": 3.9216856956481934, + "step": 2297 + }, + { + "epoch": 31.92227074235808, + "grad_norm": 0.007874754257500172, + "learning_rate": 0.0009350369459985052, + "loss": 3.879725933074951, + "step": 2298 + }, + { + "epoch": 31.936244541484715, + "grad_norm": 0.008643723092973232, + "learning_rate": 0.0009338101047467544, + "loss": 3.9303886890411377, + "step": 2299 + }, + { + "epoch": 31.950218340611354, + "grad_norm": 0.008451704867184162, + "learning_rate": 0.000932583731749205, + "loss": 3.9352827072143555, + "step": 2300 + }, + { + "epoch": 31.964192139737992, + "grad_norm": 0.008137754164636135, + "learning_rate": 0.0009313578279939236, + "loss": 3.930800437927246, + "step": 2301 + }, + { + "epoch": 31.978165938864628, + "grad_norm": 0.009277567267417908, + "learning_rate": 0.0009301323944685979, + "loss": 3.9194300174713135, + "step": 2302 + }, + { + "epoch": 31.992139737991266, + "grad_norm": 0.009428970515727997, + "learning_rate": 0.0009289074321605383, + "loss": 3.874298572540283, + "step": 2303 + }, + { + "epoch": 32.0, + "grad_norm": 0.0089727733284235, + "learning_rate": 0.0009276829420566743, + "loss": 3.893770694732666, + "step": 2304 + }, + { + "epoch": 32.0, + "eval_loss": 4.388528823852539, + "eval_runtime": 51.5859, + "eval_samples_per_second": 47.338, + "eval_steps_per_second": 1.493, + "step": 2304 + }, + { + "epoch": 32.01397379912664, + "grad_norm": 0.009470812045037746, + "learning_rate": 0.000926458925143556, + "loss": 3.8514318466186523, + "step": 2305 + }, + { + "epoch": 32.02794759825328, + "grad_norm": 0.011485518887639046, + "learning_rate": 0.0009252353824073511, + "loss": 3.857734203338623, + "step": 2306 + }, + { + "epoch": 32.041921397379916, + "grad_norm": 0.010095939971506596, + "learning_rate": 0.0009240123148338457, + "loss": 3.819117546081543, + "step": 2307 + }, + { + "epoch": 32.05589519650655, + "grad_norm": 0.010103728622198105, + "learning_rate": 0.0009227897234084444, + "loss": 3.842660427093506, + "step": 2308 + }, + { + "epoch": 32.069868995633186, + "grad_norm": 0.01148147787898779, + "learning_rate": 0.0009215676091161661, + "loss": 3.821089029312134, + "step": 2309 + }, + { + "epoch": 32.083842794759825, + "grad_norm": 0.012331553734838963, + "learning_rate": 0.0009203459729416467, + "loss": 3.8194339275360107, + "step": 2310 + }, + { + "epoch": 32.09781659388646, + "grad_norm": 0.011274794116616249, + "learning_rate": 0.0009191248158691358, + "loss": 3.8839240074157715, + "step": 2311 + }, + { + "epoch": 32.1117903930131, + "grad_norm": 0.009603365324437618, + "learning_rate": 0.0009179041388824986, + "loss": 3.806617259979248, + "step": 2312 + }, + { + "epoch": 32.12576419213974, + "grad_norm": 0.010824795812368393, + "learning_rate": 0.0009166839429652112, + "loss": 3.823228597640991, + "step": 2313 + }, + { + "epoch": 32.13973799126637, + "grad_norm": 0.01014280878007412, + "learning_rate": 0.0009154642291003637, + "loss": 3.860605239868164, + "step": 2314 + }, + { + "epoch": 32.15371179039301, + "grad_norm": 0.009052124805748463, + "learning_rate": 0.0009142449982706591, + "loss": 3.84226131439209, + "step": 2315 + }, + { + "epoch": 32.16768558951965, + "grad_norm": 0.009465130977332592, + "learning_rate": 0.0009130262514584074, + "loss": 3.8674912452697754, + "step": 2316 + }, + { + "epoch": 32.18165938864629, + "grad_norm": 0.009991783648729324, + "learning_rate": 0.0009118079896455326, + "loss": 3.8775739669799805, + "step": 2317 + }, + { + "epoch": 32.19563318777293, + "grad_norm": 0.009288253262639046, + "learning_rate": 0.0009105902138135652, + "loss": 3.7978098392486572, + "step": 2318 + }, + { + "epoch": 32.209606986899566, + "grad_norm": 0.009964323602616787, + "learning_rate": 0.0009093729249436459, + "loss": 3.8362321853637695, + "step": 2319 + }, + { + "epoch": 32.223580786026204, + "grad_norm": 0.010977157391607761, + "learning_rate": 0.0009081561240165218, + "loss": 3.857304096221924, + "step": 2320 + }, + { + "epoch": 32.237554585152836, + "grad_norm": 0.009241319261491299, + "learning_rate": 0.0009069398120125479, + "loss": 3.8659310340881348, + "step": 2321 + }, + { + "epoch": 32.251528384279474, + "grad_norm": 0.009211038239300251, + "learning_rate": 0.0009057239899116844, + "loss": 3.8764538764953613, + "step": 2322 + }, + { + "epoch": 32.26550218340611, + "grad_norm": 0.009649050422012806, + "learning_rate": 0.000904508658693497, + "loss": 3.8702778816223145, + "step": 2323 + }, + { + "epoch": 32.27947598253275, + "grad_norm": 0.00954380165785551, + "learning_rate": 0.0009032938193371573, + "loss": 3.8273327350616455, + "step": 2324 + }, + { + "epoch": 32.29344978165939, + "grad_norm": 0.008831270970404148, + "learning_rate": 0.0009020794728214378, + "loss": 3.821664810180664, + "step": 2325 + }, + { + "epoch": 32.30742358078603, + "grad_norm": 0.008367286063730717, + "learning_rate": 0.000900865620124717, + "loss": 3.8586034774780273, + "step": 2326 + }, + { + "epoch": 32.32139737991266, + "grad_norm": 0.008805174380540848, + "learning_rate": 0.0008996522622249732, + "loss": 3.811522960662842, + "step": 2327 + }, + { + "epoch": 32.3353711790393, + "grad_norm": 0.00873226672410965, + "learning_rate": 0.0008984394000997873, + "loss": 3.8538150787353516, + "step": 2328 + }, + { + "epoch": 32.34934497816594, + "grad_norm": 0.008930203504860401, + "learning_rate": 0.0008972270347263402, + "loss": 3.8448486328125, + "step": 2329 + }, + { + "epoch": 32.36331877729258, + "grad_norm": 0.009300434030592442, + "learning_rate": 0.0008960151670814125, + "loss": 3.842113494873047, + "step": 2330 + }, + { + "epoch": 32.377292576419215, + "grad_norm": 0.009598663076758385, + "learning_rate": 0.0008948037981413855, + "loss": 3.886852741241455, + "step": 2331 + }, + { + "epoch": 32.391266375545854, + "grad_norm": 0.008400370366871357, + "learning_rate": 0.0008935929288822357, + "loss": 3.852076292037964, + "step": 2332 + }, + { + "epoch": 32.40524017467249, + "grad_norm": 0.00891344528645277, + "learning_rate": 0.0008923825602795391, + "loss": 3.8480606079101562, + "step": 2333 + }, + { + "epoch": 32.419213973799124, + "grad_norm": 0.00869396049529314, + "learning_rate": 0.0008911726933084678, + "loss": 3.8164706230163574, + "step": 2334 + }, + { + "epoch": 32.43318777292576, + "grad_norm": 0.007914266549050808, + "learning_rate": 0.0008899633289437904, + "loss": 3.8768272399902344, + "step": 2335 + }, + { + "epoch": 32.4471615720524, + "grad_norm": 0.008398533798754215, + "learning_rate": 0.0008887544681598687, + "loss": 3.8625502586364746, + "step": 2336 + }, + { + "epoch": 32.46113537117904, + "grad_norm": 0.00887146033346653, + "learning_rate": 0.0008875461119306601, + "loss": 3.8934593200683594, + "step": 2337 + }, + { + "epoch": 32.47510917030568, + "grad_norm": 0.008642833679914474, + "learning_rate": 0.0008863382612297172, + "loss": 3.8391976356506348, + "step": 2338 + }, + { + "epoch": 32.48908296943232, + "grad_norm": 0.008099788799881935, + "learning_rate": 0.0008851309170301816, + "loss": 3.827965497970581, + "step": 2339 + }, + { + "epoch": 32.50305676855895, + "grad_norm": 0.008166342042386532, + "learning_rate": 0.0008839240803047894, + "loss": 3.8999152183532715, + "step": 2340 + }, + { + "epoch": 32.51703056768559, + "grad_norm": 0.009095479734241962, + "learning_rate": 0.0008827177520258671, + "loss": 3.876927614212036, + "step": 2341 + }, + { + "epoch": 32.531004366812226, + "grad_norm": 0.009246554225683212, + "learning_rate": 0.0008815119331653321, + "loss": 3.8692450523376465, + "step": 2342 + }, + { + "epoch": 32.544978165938865, + "grad_norm": 0.007059223018586636, + "learning_rate": 0.0008803066246946898, + "loss": 3.869400978088379, + "step": 2343 + }, + { + "epoch": 32.5589519650655, + "grad_norm": 0.008499744348227978, + "learning_rate": 0.0008791018275850363, + "loss": 3.8303189277648926, + "step": 2344 + }, + { + "epoch": 32.57292576419214, + "grad_norm": 0.007667839527130127, + "learning_rate": 0.0008778975428070543, + "loss": 3.827998638153076, + "step": 2345 + }, + { + "epoch": 32.58689956331878, + "grad_norm": 0.00765089550986886, + "learning_rate": 0.0008766937713310145, + "loss": 3.8554482460021973, + "step": 2346 + }, + { + "epoch": 32.60087336244541, + "grad_norm": 0.00862938817590475, + "learning_rate": 0.0008754905141267749, + "loss": 3.905458927154541, + "step": 2347 + }, + { + "epoch": 32.61484716157205, + "grad_norm": 0.008733712136745453, + "learning_rate": 0.0008742877721637758, + "loss": 3.8564133644104004, + "step": 2348 + }, + { + "epoch": 32.62882096069869, + "grad_norm": 0.0071738120168447495, + "learning_rate": 0.0008730855464110471, + "loss": 3.878807544708252, + "step": 2349 + }, + { + "epoch": 32.64279475982533, + "grad_norm": 0.008418241515755653, + "learning_rate": 0.0008718838378371989, + "loss": 3.8399720191955566, + "step": 2350 + }, + { + "epoch": 32.65676855895197, + "grad_norm": 0.008526502177119255, + "learning_rate": 0.0008706826474104263, + "loss": 3.847656488418579, + "step": 2351 + }, + { + "epoch": 32.670742358078606, + "grad_norm": 0.00732462527230382, + "learning_rate": 0.0008694819760985069, + "loss": 3.8839492797851562, + "step": 2352 + }, + { + "epoch": 32.68471615720524, + "grad_norm": 0.008424196392297745, + "learning_rate": 0.0008682818248688003, + "loss": 3.861239433288574, + "step": 2353 + }, + { + "epoch": 32.698689956331876, + "grad_norm": 0.008984955959022045, + "learning_rate": 0.0008670821946882465, + "loss": 3.856529712677002, + "step": 2354 + }, + { + "epoch": 32.712663755458514, + "grad_norm": 0.009004034101963043, + "learning_rate": 0.000865883086523365, + "loss": 3.8514528274536133, + "step": 2355 + }, + { + "epoch": 32.72663755458515, + "grad_norm": 0.009097260423004627, + "learning_rate": 0.0008646845013402563, + "loss": 3.8890581130981445, + "step": 2356 + }, + { + "epoch": 32.74061135371179, + "grad_norm": 0.009244748391211033, + "learning_rate": 0.0008634864401045988, + "loss": 3.872745990753174, + "step": 2357 + }, + { + "epoch": 32.75458515283843, + "grad_norm": 0.00788815412670374, + "learning_rate": 0.0008622889037816488, + "loss": 3.8777575492858887, + "step": 2358 + }, + { + "epoch": 32.76855895196506, + "grad_norm": 0.008943610824644566, + "learning_rate": 0.0008610918933362396, + "loss": 3.8825736045837402, + "step": 2359 + }, + { + "epoch": 32.7825327510917, + "grad_norm": 0.008154108189046383, + "learning_rate": 0.0008598954097327809, + "loss": 3.877514362335205, + "step": 2360 + }, + { + "epoch": 32.79650655021834, + "grad_norm": 0.008689315989613533, + "learning_rate": 0.0008586994539352587, + "loss": 3.854720115661621, + "step": 2361 + }, + { + "epoch": 32.81048034934498, + "grad_norm": 0.00904505793005228, + "learning_rate": 0.0008575040269072315, + "loss": 3.911463499069214, + "step": 2362 + }, + { + "epoch": 32.82445414847162, + "grad_norm": 0.007773210760205984, + "learning_rate": 0.0008563091296118342, + "loss": 3.875370740890503, + "step": 2363 + }, + { + "epoch": 32.838427947598255, + "grad_norm": 0.008437697775661945, + "learning_rate": 0.0008551147630117737, + "loss": 3.8457469940185547, + "step": 2364 + }, + { + "epoch": 32.852401746724894, + "grad_norm": 0.009780957363545895, + "learning_rate": 0.0008539209280693296, + "loss": 3.890479564666748, + "step": 2365 + }, + { + "epoch": 32.866375545851525, + "grad_norm": 0.007929213345050812, + "learning_rate": 0.0008527276257463532, + "loss": 3.876030445098877, + "step": 2366 + }, + { + "epoch": 32.880349344978164, + "grad_norm": 0.009121386334300041, + "learning_rate": 0.0008515348570042669, + "loss": 3.9252734184265137, + "step": 2367 + }, + { + "epoch": 32.8943231441048, + "grad_norm": 0.00981531199067831, + "learning_rate": 0.000850342622804063, + "loss": 3.8997983932495117, + "step": 2368 + }, + { + "epoch": 32.90829694323144, + "grad_norm": 0.009150844067335129, + "learning_rate": 0.0008491509241063025, + "loss": 3.85115385055542, + "step": 2369 + }, + { + "epoch": 32.92227074235808, + "grad_norm": 0.00992978923022747, + "learning_rate": 0.0008479597618711154, + "loss": 3.9138102531433105, + "step": 2370 + }, + { + "epoch": 32.93624454148472, + "grad_norm": 0.010906689800322056, + "learning_rate": 0.0008467691370582001, + "loss": 3.892329692840576, + "step": 2371 + }, + { + "epoch": 32.95021834061135, + "grad_norm": 0.008644519373774529, + "learning_rate": 0.0008455790506268213, + "loss": 3.9071311950683594, + "step": 2372 + }, + { + "epoch": 32.96419213973799, + "grad_norm": 0.009030954912304878, + "learning_rate": 0.0008443895035358111, + "loss": 3.867384910583496, + "step": 2373 + }, + { + "epoch": 32.97816593886463, + "grad_norm": 0.010806960985064507, + "learning_rate": 0.0008432004967435638, + "loss": 3.877983331680298, + "step": 2374 + }, + { + "epoch": 32.992139737991266, + "grad_norm": 0.011148087680339813, + "learning_rate": 0.0008420120312080436, + "loss": 3.8795924186706543, + "step": 2375 + }, + { + "epoch": 33.0, + "grad_norm": 0.01046178862452507, + "learning_rate": 0.0008408241078867731, + "loss": 3.887709617614746, + "step": 2376 + }, + { + "epoch": 33.0, + "eval_loss": 4.383835792541504, + "eval_runtime": 53.0778, + "eval_samples_per_second": 46.008, + "eval_steps_per_second": 1.451, + "step": 2376 + }, + { + "epoch": 33.01397379912664, + "grad_norm": 0.009721867740154266, + "learning_rate": 0.0008396367277368421, + "loss": 3.8261327743530273, + "step": 2377 + }, + { + "epoch": 33.02794759825328, + "grad_norm": 0.01091420091688633, + "learning_rate": 0.000838449891714901, + "loss": 3.7837276458740234, + "step": 2378 + }, + { + "epoch": 33.041921397379916, + "grad_norm": 0.010423852130770683, + "learning_rate": 0.0008372636007771623, + "loss": 3.796450138092041, + "step": 2379 + }, + { + "epoch": 33.05589519650655, + "grad_norm": 0.011928882449865341, + "learning_rate": 0.0008360778558793996, + "loss": 3.794717311859131, + "step": 2380 + }, + { + "epoch": 33.069868995633186, + "grad_norm": 0.010779444128274918, + "learning_rate": 0.000834892657976945, + "loss": 3.8477983474731445, + "step": 2381 + }, + { + "epoch": 33.083842794759825, + "grad_norm": 0.009554282762110233, + "learning_rate": 0.0008337080080246915, + "loss": 3.793790817260742, + "step": 2382 + }, + { + "epoch": 33.09781659388646, + "grad_norm": 0.009768152609467506, + "learning_rate": 0.0008325239069770906, + "loss": 3.860384225845337, + "step": 2383 + }, + { + "epoch": 33.1117903930131, + "grad_norm": 0.010801832191646099, + "learning_rate": 0.0008313403557881513, + "loss": 3.816796064376831, + "step": 2384 + }, + { + "epoch": 33.12576419213974, + "grad_norm": 0.01046004518866539, + "learning_rate": 0.0008301573554114378, + "loss": 3.8064301013946533, + "step": 2385 + }, + { + "epoch": 33.13973799126637, + "grad_norm": 0.009319447912275791, + "learning_rate": 0.0008289749068000744, + "loss": 3.832408905029297, + "step": 2386 + }, + { + "epoch": 33.15371179039301, + "grad_norm": 0.01011407095938921, + "learning_rate": 0.000827793010906737, + "loss": 3.8276009559631348, + "step": 2387 + }, + { + "epoch": 33.16768558951965, + "grad_norm": 0.010393989272415638, + "learning_rate": 0.0008266116686836582, + "loss": 3.83754825592041, + "step": 2388 + }, + { + "epoch": 33.18165938864629, + "grad_norm": 0.00883004255592823, + "learning_rate": 0.0008254308810826244, + "loss": 3.8315513134002686, + "step": 2389 + }, + { + "epoch": 33.19563318777293, + "grad_norm": 0.008549787104129791, + "learning_rate": 0.0008242506490549745, + "loss": 3.792919158935547, + "step": 2390 + }, + { + "epoch": 33.209606986899566, + "grad_norm": 0.009668390266597271, + "learning_rate": 0.0008230709735516006, + "loss": 3.8110954761505127, + "step": 2391 + }, + { + "epoch": 33.223580786026204, + "grad_norm": 0.009008699096739292, + "learning_rate": 0.0008218918555229445, + "loss": 3.8598508834838867, + "step": 2392 + }, + { + "epoch": 33.237554585152836, + "grad_norm": 0.008660725317895412, + "learning_rate": 0.0008207132959190026, + "loss": 3.839632511138916, + "step": 2393 + }, + { + "epoch": 33.251528384279474, + "grad_norm": 0.00827972125262022, + "learning_rate": 0.0008195352956893172, + "loss": 3.8428683280944824, + "step": 2394 + }, + { + "epoch": 33.26550218340611, + "grad_norm": 0.008117233403027058, + "learning_rate": 0.0008183578557829825, + "loss": 3.8434956073760986, + "step": 2395 + }, + { + "epoch": 33.27947598253275, + "grad_norm": 0.008314455859363079, + "learning_rate": 0.0008171809771486405, + "loss": 3.8079686164855957, + "step": 2396 + }, + { + "epoch": 33.29344978165939, + "grad_norm": 0.00865850131958723, + "learning_rate": 0.0008160046607344811, + "loss": 3.846999406814575, + "step": 2397 + }, + { + "epoch": 33.30742358078603, + "grad_norm": 0.0096864253282547, + "learning_rate": 0.0008148289074882416, + "loss": 3.8227243423461914, + "step": 2398 + }, + { + "epoch": 33.32139737991266, + "grad_norm": 0.00840236246585846, + "learning_rate": 0.0008136537183572045, + "loss": 3.8400344848632812, + "step": 2399 + }, + { + "epoch": 33.3353711790393, + "grad_norm": 0.007451791316270828, + "learning_rate": 0.0008124790942881984, + "loss": 3.8318095207214355, + "step": 2400 + }, + { + "epoch": 33.34934497816594, + "grad_norm": 0.00858070608228445, + "learning_rate": 0.000811305036227597, + "loss": 3.8232922554016113, + "step": 2401 + }, + { + "epoch": 33.36331877729258, + "grad_norm": 0.00765061192214489, + "learning_rate": 0.0008101315451213178, + "loss": 3.81624174118042, + "step": 2402 + }, + { + "epoch": 33.377292576419215, + "grad_norm": 0.009304090403020382, + "learning_rate": 0.0008089586219148213, + "loss": 3.8715004920959473, + "step": 2403 + }, + { + "epoch": 33.391266375545854, + "grad_norm": 0.009136030450463295, + "learning_rate": 0.0008077862675531103, + "loss": 3.825037956237793, + "step": 2404 + }, + { + "epoch": 33.40524017467249, + "grad_norm": 0.007460939697921276, + "learning_rate": 0.0008066144829807303, + "loss": 3.811749219894409, + "step": 2405 + }, + { + "epoch": 33.419213973799124, + "grad_norm": 0.007510480936616659, + "learning_rate": 0.0008054432691417659, + "loss": 3.8313679695129395, + "step": 2406 + }, + { + "epoch": 33.43318777292576, + "grad_norm": 0.008825462311506271, + "learning_rate": 0.0008042726269798433, + "loss": 3.83119535446167, + "step": 2407 + }, + { + "epoch": 33.4471615720524, + "grad_norm": 0.009002318605780602, + "learning_rate": 0.0008031025574381281, + "loss": 3.8542468547821045, + "step": 2408 + }, + { + "epoch": 33.46113537117904, + "grad_norm": 0.009266122244298458, + "learning_rate": 0.0008019330614593237, + "loss": 3.8328065872192383, + "step": 2409 + }, + { + "epoch": 33.47510917030568, + "grad_norm": 0.008935115300118923, + "learning_rate": 0.0008007641399856727, + "loss": 3.8198719024658203, + "step": 2410 + }, + { + "epoch": 33.48908296943232, + "grad_norm": 0.008021850138902664, + "learning_rate": 0.0007995957939589526, + "loss": 3.828382968902588, + "step": 2411 + }, + { + "epoch": 33.50305676855895, + "grad_norm": 0.008035171777009964, + "learning_rate": 0.0007984280243204805, + "loss": 3.879868507385254, + "step": 2412 + }, + { + "epoch": 33.51703056768559, + "grad_norm": 0.008553579449653625, + "learning_rate": 0.0007972608320111059, + "loss": 3.842162609100342, + "step": 2413 + }, + { + "epoch": 33.531004366812226, + "grad_norm": 0.008016746491193771, + "learning_rate": 0.000796094217971215, + "loss": 3.85674786567688, + "step": 2414 + }, + { + "epoch": 33.544978165938865, + "grad_norm": 0.007781418971717358, + "learning_rate": 0.0007949281831407276, + "loss": 3.7968859672546387, + "step": 2415 + }, + { + "epoch": 33.5589519650655, + "grad_norm": 0.008552255108952522, + "learning_rate": 0.0007937627284590971, + "loss": 3.8305506706237793, + "step": 2416 + }, + { + "epoch": 33.57292576419214, + "grad_norm": 0.007390565704554319, + "learning_rate": 0.0007925978548653096, + "loss": 3.8228843212127686, + "step": 2417 + }, + { + "epoch": 33.58689956331878, + "grad_norm": 0.007611850742250681, + "learning_rate": 0.0007914335632978812, + "loss": 3.868957996368408, + "step": 2418 + }, + { + "epoch": 33.60087336244541, + "grad_norm": 0.008188549429178238, + "learning_rate": 0.0007902698546948628, + "loss": 3.836078405380249, + "step": 2419 + }, + { + "epoch": 33.61484716157205, + "grad_norm": 0.008198797702789307, + "learning_rate": 0.0007891067299938318, + "loss": 3.8637044429779053, + "step": 2420 + }, + { + "epoch": 33.62882096069869, + "grad_norm": 0.008353965356945992, + "learning_rate": 0.0007879441901318977, + "loss": 3.833843946456909, + "step": 2421 + }, + { + "epoch": 33.64279475982533, + "grad_norm": 0.0086967209354043, + "learning_rate": 0.0007867822360456963, + "loss": 3.850536823272705, + "step": 2422 + }, + { + "epoch": 33.65676855895197, + "grad_norm": 0.009151408448815346, + "learning_rate": 0.0007856208686713949, + "loss": 3.8453407287597656, + "step": 2423 + }, + { + "epoch": 33.670742358078606, + "grad_norm": 0.008010942488908768, + "learning_rate": 0.0007844600889446858, + "loss": 3.8679513931274414, + "step": 2424 + }, + { + "epoch": 33.68471615720524, + "grad_norm": 0.007813206873834133, + "learning_rate": 0.0007832998978007873, + "loss": 3.860609531402588, + "step": 2425 + }, + { + "epoch": 33.698689956331876, + "grad_norm": 0.007759213913232088, + "learning_rate": 0.0007821402961744448, + "loss": 3.803863048553467, + "step": 2426 + }, + { + "epoch": 33.712663755458514, + "grad_norm": 0.008045864291489124, + "learning_rate": 0.0007809812849999293, + "loss": 3.858161449432373, + "step": 2427 + }, + { + "epoch": 33.72663755458515, + "grad_norm": 0.008212368935346603, + "learning_rate": 0.0007798228652110347, + "loss": 3.8311336040496826, + "step": 2428 + }, + { + "epoch": 33.74061135371179, + "grad_norm": 0.007982932962477207, + "learning_rate": 0.0007786650377410774, + "loss": 3.870339870452881, + "step": 2429 + }, + { + "epoch": 33.75458515283843, + "grad_norm": 0.0071402136236429214, + "learning_rate": 0.0007775078035229002, + "loss": 3.7922263145446777, + "step": 2430 + }, + { + "epoch": 33.76855895196506, + "grad_norm": 0.008564750663936138, + "learning_rate": 0.0007763511634888656, + "loss": 3.83981990814209, + "step": 2431 + }, + { + "epoch": 33.7825327510917, + "grad_norm": 0.008078585378825665, + "learning_rate": 0.0007751951185708562, + "loss": 3.852938652038574, + "step": 2432 + }, + { + "epoch": 33.79650655021834, + "grad_norm": 0.00801754929125309, + "learning_rate": 0.0007740396697002774, + "loss": 3.8434009552001953, + "step": 2433 + }, + { + "epoch": 33.81048034934498, + "grad_norm": 0.008973612450063229, + "learning_rate": 0.0007728848178080535, + "loss": 3.8238563537597656, + "step": 2434 + }, + { + "epoch": 33.82445414847162, + "grad_norm": 0.010199865326285362, + "learning_rate": 0.0007717305638246281, + "loss": 3.8891735076904297, + "step": 2435 + }, + { + "epoch": 33.838427947598255, + "grad_norm": 0.009620471857488155, + "learning_rate": 0.0007705769086799617, + "loss": 3.8611769676208496, + "step": 2436 + }, + { + "epoch": 33.852401746724894, + "grad_norm": 0.007981563918292522, + "learning_rate": 0.0007694238533035337, + "loss": 3.8496928215026855, + "step": 2437 + }, + { + "epoch": 33.866375545851525, + "grad_norm": 0.008856567554175854, + "learning_rate": 0.0007682713986243415, + "loss": 3.8366098403930664, + "step": 2438 + }, + { + "epoch": 33.880349344978164, + "grad_norm": 0.008323253132402897, + "learning_rate": 0.0007671195455708952, + "loss": 3.8224244117736816, + "step": 2439 + }, + { + "epoch": 33.8943231441048, + "grad_norm": 0.00814893003553152, + "learning_rate": 0.0007659682950712228, + "loss": 3.8611154556274414, + "step": 2440 + }, + { + "epoch": 33.90829694323144, + "grad_norm": 0.009146672673523426, + "learning_rate": 0.0007648176480528659, + "loss": 3.8258843421936035, + "step": 2441 + }, + { + "epoch": 33.92227074235808, + "grad_norm": 0.007573423441499472, + "learning_rate": 0.0007636676054428805, + "loss": 3.888007640838623, + "step": 2442 + }, + { + "epoch": 33.93624454148472, + "grad_norm": 0.00813984964042902, + "learning_rate": 0.0007625181681678343, + "loss": 3.860306739807129, + "step": 2443 + }, + { + "epoch": 33.95021834061135, + "grad_norm": 0.007936778478324413, + "learning_rate": 0.0007613693371538078, + "loss": 3.8808555603027344, + "step": 2444 + }, + { + "epoch": 33.96419213973799, + "grad_norm": 0.008434669114649296, + "learning_rate": 0.0007602211133263956, + "loss": 3.854154348373413, + "step": 2445 + }, + { + "epoch": 33.97816593886463, + "grad_norm": 0.007857564836740494, + "learning_rate": 0.0007590734976106985, + "loss": 3.8240737915039062, + "step": 2446 + }, + { + "epoch": 33.992139737991266, + "grad_norm": 0.007989327423274517, + "learning_rate": 0.0007579264909313315, + "loss": 3.850971221923828, + "step": 2447 + }, + { + "epoch": 34.0, + "grad_norm": 0.008888393640518188, + "learning_rate": 0.0007567800942124152, + "loss": 3.862267255783081, + "step": 2448 + }, + { + "epoch": 34.0, + "eval_loss": 4.38022518157959, + "eval_runtime": 51.4686, + "eval_samples_per_second": 47.446, + "eval_steps_per_second": 1.496, + "step": 2448 + }, + { + "epoch": 34.01397379912664, + "grad_norm": 0.00952722318470478, + "learning_rate": 0.0007556343083775832, + "loss": 3.7705495357513428, + "step": 2449 + }, + { + "epoch": 34.02794759825328, + "grad_norm": 0.01099323108792305, + "learning_rate": 0.0007544891343499723, + "loss": 3.80812406539917, + "step": 2450 + }, + { + "epoch": 34.041921397379916, + "grad_norm": 0.010298571549355984, + "learning_rate": 0.0007533445730522295, + "loss": 3.782794237136841, + "step": 2451 + }, + { + "epoch": 34.05589519650655, + "grad_norm": 0.008139589801430702, + "learning_rate": 0.000752200625406507, + "loss": 3.7877392768859863, + "step": 2452 + }, + { + "epoch": 34.069868995633186, + "grad_norm": 0.009031252935528755, + "learning_rate": 0.0007510572923344628, + "loss": 3.8181564807891846, + "step": 2453 + }, + { + "epoch": 34.083842794759825, + "grad_norm": 0.00942459236830473, + "learning_rate": 0.0007499145747572603, + "loss": 3.807425022125244, + "step": 2454 + }, + { + "epoch": 34.09781659388646, + "grad_norm": 0.00922947097569704, + "learning_rate": 0.0007487724735955648, + "loss": 3.7923264503479004, + "step": 2455 + }, + { + "epoch": 34.1117903930131, + "grad_norm": 0.009551790542900562, + "learning_rate": 0.0007476309897695485, + "loss": 3.790083885192871, + "step": 2456 + }, + { + "epoch": 34.12576419213974, + "grad_norm": 0.008996761403977871, + "learning_rate": 0.0007464901241988831, + "loss": 3.836491107940674, + "step": 2457 + }, + { + "epoch": 34.13973799126637, + "grad_norm": 0.009766182862222195, + "learning_rate": 0.0007453498778027433, + "loss": 3.7829647064208984, + "step": 2458 + }, + { + "epoch": 34.15371179039301, + "grad_norm": 0.007708575110882521, + "learning_rate": 0.0007442102514998058, + "loss": 3.803400993347168, + "step": 2459 + }, + { + "epoch": 34.16768558951965, + "grad_norm": 0.008606902323663235, + "learning_rate": 0.0007430712462082464, + "loss": 3.810825824737549, + "step": 2460 + }, + { + "epoch": 34.18165938864629, + "grad_norm": 0.009462927468121052, + "learning_rate": 0.0007419328628457416, + "loss": 3.784748077392578, + "step": 2461 + }, + { + "epoch": 34.19563318777293, + "grad_norm": 0.010283366777002811, + "learning_rate": 0.0007407951023294655, + "loss": 3.744312286376953, + "step": 2462 + }, + { + "epoch": 34.209606986899566, + "grad_norm": 0.008952794596552849, + "learning_rate": 0.0007396579655760917, + "loss": 3.8098015785217285, + "step": 2463 + }, + { + "epoch": 34.223580786026204, + "grad_norm": 0.008446207270026207, + "learning_rate": 0.0007385214535017907, + "loss": 3.8306760787963867, + "step": 2464 + }, + { + "epoch": 34.237554585152836, + "grad_norm": 0.007988285273313522, + "learning_rate": 0.0007373855670222302, + "loss": 3.7976696491241455, + "step": 2465 + }, + { + "epoch": 34.251528384279474, + "grad_norm": 0.007586437277495861, + "learning_rate": 0.0007362503070525718, + "loss": 3.785891056060791, + "step": 2466 + }, + { + "epoch": 34.26550218340611, + "grad_norm": 0.00775605533272028, + "learning_rate": 0.0007351156745074757, + "loss": 3.8176379203796387, + "step": 2467 + }, + { + "epoch": 34.27947598253275, + "grad_norm": 0.007727691903710365, + "learning_rate": 0.0007339816703010948, + "loss": 3.791743278503418, + "step": 2468 + }, + { + "epoch": 34.29344978165939, + "grad_norm": 0.008311517536640167, + "learning_rate": 0.0007328482953470747, + "loss": 3.818769693374634, + "step": 2469 + }, + { + "epoch": 34.30742358078603, + "grad_norm": 0.007610874716192484, + "learning_rate": 0.0007317155505585556, + "loss": 3.76772141456604, + "step": 2470 + }, + { + "epoch": 34.32139737991266, + "grad_norm": 0.0078107016161084175, + "learning_rate": 0.0007305834368481699, + "loss": 3.822758197784424, + "step": 2471 + }, + { + "epoch": 34.3353711790393, + "grad_norm": 0.006906864698976278, + "learning_rate": 0.0007294519551280412, + "loss": 3.832834005355835, + "step": 2472 + }, + { + "epoch": 34.34934497816594, + "grad_norm": 0.008230660110712051, + "learning_rate": 0.0007283211063097829, + "loss": 3.85087513923645, + "step": 2473 + }, + { + "epoch": 34.36331877729258, + "grad_norm": 0.008201118558645248, + "learning_rate": 0.0007271908913044997, + "loss": 3.7777247428894043, + "step": 2474 + }, + { + "epoch": 34.377292576419215, + "grad_norm": 0.0070146857760846615, + "learning_rate": 0.000726061311022787, + "loss": 3.837517738342285, + "step": 2475 + }, + { + "epoch": 34.391266375545854, + "grad_norm": 0.0076036700047552586, + "learning_rate": 0.0007249323663747254, + "loss": 3.8358592987060547, + "step": 2476 + }, + { + "epoch": 34.40524017467249, + "grad_norm": 0.007238952908664942, + "learning_rate": 0.0007238040582698858, + "loss": 3.828089952468872, + "step": 2477 + }, + { + "epoch": 34.419213973799124, + "grad_norm": 0.00825269054621458, + "learning_rate": 0.0007226763876173259, + "loss": 3.806016445159912, + "step": 2478 + }, + { + "epoch": 34.43318777292576, + "grad_norm": 0.008368568494915962, + "learning_rate": 0.0007215493553255898, + "loss": 3.8163394927978516, + "step": 2479 + }, + { + "epoch": 34.4471615720524, + "grad_norm": 0.00728471577167511, + "learning_rate": 0.0007204229623027063, + "loss": 3.835597038269043, + "step": 2480 + }, + { + "epoch": 34.46113537117904, + "grad_norm": 0.00784830842167139, + "learning_rate": 0.0007192972094561897, + "loss": 3.8050971031188965, + "step": 2481 + }, + { + "epoch": 34.47510917030568, + "grad_norm": 0.007714141625910997, + "learning_rate": 0.0007181720976930404, + "loss": 3.8031606674194336, + "step": 2482 + }, + { + "epoch": 34.48908296943232, + "grad_norm": 0.009128103032708168, + "learning_rate": 0.000717047627919739, + "loss": 3.77107572555542, + "step": 2483 + }, + { + "epoch": 34.50305676855895, + "grad_norm": 0.008033272810280323, + "learning_rate": 0.000715923801042251, + "loss": 3.8615059852600098, + "step": 2484 + }, + { + "epoch": 34.51703056768559, + "grad_norm": 0.007732565049082041, + "learning_rate": 0.0007148006179660235, + "loss": 3.8277132511138916, + "step": 2485 + }, + { + "epoch": 34.531004366812226, + "grad_norm": 0.007382966112345457, + "learning_rate": 0.0007136780795959853, + "loss": 3.803431987762451, + "step": 2486 + }, + { + "epoch": 34.544978165938865, + "grad_norm": 0.007517020218074322, + "learning_rate": 0.0007125561868365442, + "loss": 3.799734354019165, + "step": 2487 + }, + { + "epoch": 34.5589519650655, + "grad_norm": 0.008041434921324253, + "learning_rate": 0.0007114349405915895, + "loss": 3.7848258018493652, + "step": 2488 + }, + { + "epoch": 34.57292576419214, + "grad_norm": 0.007875649258494377, + "learning_rate": 0.0007103143417644891, + "loss": 3.8136563301086426, + "step": 2489 + }, + { + "epoch": 34.58689956331878, + "grad_norm": 0.008213640190660954, + "learning_rate": 0.000709194391258089, + "loss": 3.8518495559692383, + "step": 2490 + }, + { + "epoch": 34.60087336244541, + "grad_norm": 0.00801054760813713, + "learning_rate": 0.0007080750899747139, + "loss": 3.7956271171569824, + "step": 2491 + }, + { + "epoch": 34.61484716157205, + "grad_norm": 0.008064007386565208, + "learning_rate": 0.0007069564388161631, + "loss": 3.854595184326172, + "step": 2492 + }, + { + "epoch": 34.62882096069869, + "grad_norm": 0.008300521411001682, + "learning_rate": 0.0007058384386837154, + "loss": 3.8203625679016113, + "step": 2493 + }, + { + "epoch": 34.64279475982533, + "grad_norm": 0.008640496991574764, + "learning_rate": 0.000704721090478122, + "loss": 3.793574810028076, + "step": 2494 + }, + { + "epoch": 34.65676855895197, + "grad_norm": 0.007684200536459684, + "learning_rate": 0.0007036043950996105, + "loss": 3.8564438819885254, + "step": 2495 + }, + { + "epoch": 34.670742358078606, + "grad_norm": 0.008431905880570412, + "learning_rate": 0.0007024883534478823, + "loss": 3.8271119594573975, + "step": 2496 + }, + { + "epoch": 34.68471615720524, + "grad_norm": 0.008137186989188194, + "learning_rate": 0.0007013729664221122, + "loss": 3.8219337463378906, + "step": 2497 + }, + { + "epoch": 34.698689956331876, + "grad_norm": 0.0076744635589420795, + "learning_rate": 0.0007002582349209474, + "loss": 3.8353049755096436, + "step": 2498 + }, + { + "epoch": 34.712663755458514, + "grad_norm": 0.008388740941882133, + "learning_rate": 0.0006991441598425065, + "loss": 3.795104503631592, + "step": 2499 + }, + { + "epoch": 34.72663755458515, + "grad_norm": 0.007856911979615688, + "learning_rate": 0.0006980307420843798, + "loss": 3.818411111831665, + "step": 2500 + }, + { + "epoch": 34.74061135371179, + "grad_norm": 0.008048887364566326, + "learning_rate": 0.000696917982543628, + "loss": 3.86599063873291, + "step": 2501 + }, + { + "epoch": 34.75458515283843, + "grad_norm": 0.007949979975819588, + "learning_rate": 0.0006958058821167812, + "loss": 3.8026742935180664, + "step": 2502 + }, + { + "epoch": 34.76855895196506, + "grad_norm": 0.008162821643054485, + "learning_rate": 0.0006946944416998389, + "loss": 3.833751678466797, + "step": 2503 + }, + { + "epoch": 34.7825327510917, + "grad_norm": 0.008190765976905823, + "learning_rate": 0.0006935836621882682, + "loss": 3.8164236545562744, + "step": 2504 + }, + { + "epoch": 34.79650655021834, + "grad_norm": 0.009714310988783836, + "learning_rate": 0.000692473544477005, + "loss": 3.8367366790771484, + "step": 2505 + }, + { + "epoch": 34.81048034934498, + "grad_norm": 0.008075987920165062, + "learning_rate": 0.0006913640894604498, + "loss": 3.8108773231506348, + "step": 2506 + }, + { + "epoch": 34.82445414847162, + "grad_norm": 0.009075167588889599, + "learning_rate": 0.0006902552980324713, + "loss": 3.8444485664367676, + "step": 2507 + }, + { + "epoch": 34.838427947598255, + "grad_norm": 0.008678440935909748, + "learning_rate": 0.0006891471710864022, + "loss": 3.8384461402893066, + "step": 2508 + }, + { + "epoch": 34.852401746724894, + "grad_norm": 0.007705411408096552, + "learning_rate": 0.0006880397095150414, + "loss": 3.825443744659424, + "step": 2509 + }, + { + "epoch": 34.866375545851525, + "grad_norm": 0.00782099086791277, + "learning_rate": 0.0006869329142106488, + "loss": 3.8331687450408936, + "step": 2510 + }, + { + "epoch": 34.880349344978164, + "grad_norm": 0.007542501203715801, + "learning_rate": 0.0006858267860649509, + "loss": 3.7905991077423096, + "step": 2511 + }, + { + "epoch": 34.8943231441048, + "grad_norm": 0.008315361104905605, + "learning_rate": 0.0006847213259691356, + "loss": 3.861527442932129, + "step": 2512 + }, + { + "epoch": 34.90829694323144, + "grad_norm": 0.008529976941645145, + "learning_rate": 0.0006836165348138508, + "loss": 3.8332972526550293, + "step": 2513 + }, + { + "epoch": 34.92227074235808, + "grad_norm": 0.008117211051285267, + "learning_rate": 0.0006825124134892072, + "loss": 3.7961089611053467, + "step": 2514 + }, + { + "epoch": 34.93624454148472, + "grad_norm": 0.008409497328102589, + "learning_rate": 0.0006814089628847755, + "loss": 3.812218189239502, + "step": 2515 + }, + { + "epoch": 34.95021834061135, + "grad_norm": 0.007592010777443647, + "learning_rate": 0.0006803061838895864, + "loss": 3.801304340362549, + "step": 2516 + }, + { + "epoch": 34.96419213973799, + "grad_norm": 0.007820053026080132, + "learning_rate": 0.0006792040773921283, + "loss": 3.8517117500305176, + "step": 2517 + }, + { + "epoch": 34.97816593886463, + "grad_norm": 0.0073044621385633945, + "learning_rate": 0.0006781026442803479, + "loss": 3.8253560066223145, + "step": 2518 + }, + { + "epoch": 34.992139737991266, + "grad_norm": 0.007546546868979931, + "learning_rate": 0.0006770018854416522, + "loss": 3.800802230834961, + "step": 2519 + }, + { + "epoch": 35.0, + "grad_norm": 0.008878474123775959, + "learning_rate": 0.0006759018017629007, + "loss": 3.8133859634399414, + "step": 2520 + }, + { + "epoch": 35.0, + "eval_loss": 4.375829219818115, + "eval_runtime": 51.3756, + "eval_samples_per_second": 47.532, + "eval_steps_per_second": 1.499, + "step": 2520 + }, + { + "epoch": 35.01397379912664, + "grad_norm": 0.00814585480839014, + "learning_rate": 0.0006748023941304115, + "loss": 3.7970919609069824, + "step": 2521 + }, + { + "epoch": 35.02794759825328, + "grad_norm": 0.00824262760579586, + "learning_rate": 0.0006737036634299577, + "loss": 3.750131607055664, + "step": 2522 + }, + { + "epoch": 35.041921397379916, + "grad_norm": 0.008675593882799149, + "learning_rate": 0.000672605610546767, + "loss": 3.791456699371338, + "step": 2523 + }, + { + "epoch": 35.05589519650655, + "grad_norm": 0.009052547626197338, + "learning_rate": 0.00067150823636552, + "loss": 3.7762060165405273, + "step": 2524 + }, + { + "epoch": 35.069868995633186, + "grad_norm": 0.00818016566336155, + "learning_rate": 0.0006704115417703514, + "loss": 3.750631332397461, + "step": 2525 + }, + { + "epoch": 35.083842794759825, + "grad_norm": 0.00800115056335926, + "learning_rate": 0.0006693155276448489, + "loss": 3.811519145965576, + "step": 2526 + }, + { + "epoch": 35.09781659388646, + "grad_norm": 0.009032433852553368, + "learning_rate": 0.0006682201948720509, + "loss": 3.7805609703063965, + "step": 2527 + }, + { + "epoch": 35.1117903930131, + "grad_norm": 0.0077886078506708145, + "learning_rate": 0.0006671255443344474, + "loss": 3.749366521835327, + "step": 2528 + }, + { + "epoch": 35.12576419213974, + "grad_norm": 0.008603204973042011, + "learning_rate": 0.0006660315769139774, + "loss": 3.7686312198638916, + "step": 2529 + }, + { + "epoch": 35.13973799126637, + "grad_norm": 0.008361323736608028, + "learning_rate": 0.0006649382934920329, + "loss": 3.740269660949707, + "step": 2530 + }, + { + "epoch": 35.15371179039301, + "grad_norm": 0.00948982872068882, + "learning_rate": 0.0006638456949494507, + "loss": 3.7874045372009277, + "step": 2531 + }, + { + "epoch": 35.16768558951965, + "grad_norm": 0.008958354592323303, + "learning_rate": 0.0006627537821665185, + "loss": 3.78987455368042, + "step": 2532 + }, + { + "epoch": 35.18165938864629, + "grad_norm": 0.008702470920979977, + "learning_rate": 0.0006616625560229704, + "loss": 3.751061201095581, + "step": 2533 + }, + { + "epoch": 35.19563318777293, + "grad_norm": 0.008424391970038414, + "learning_rate": 0.000660572017397988, + "loss": 3.80816650390625, + "step": 2534 + }, + { + "epoch": 35.209606986899566, + "grad_norm": 0.009022071026265621, + "learning_rate": 0.0006594821671701989, + "loss": 3.726067304611206, + "step": 2535 + }, + { + "epoch": 35.223580786026204, + "grad_norm": 0.007974586449563503, + "learning_rate": 0.0006583930062176739, + "loss": 3.78564453125, + "step": 2536 + }, + { + "epoch": 35.237554585152836, + "grad_norm": 0.007887622341513634, + "learning_rate": 0.0006573045354179329, + "loss": 3.7993059158325195, + "step": 2537 + }, + { + "epoch": 35.251528384279474, + "grad_norm": 0.007782666943967342, + "learning_rate": 0.0006562167556479352, + "loss": 3.7440507411956787, + "step": 2538 + }, + { + "epoch": 35.26550218340611, + "grad_norm": 0.007923880591988564, + "learning_rate": 0.000655129667784086, + "loss": 3.7920784950256348, + "step": 2539 + }, + { + "epoch": 35.27947598253275, + "grad_norm": 0.007850701920688152, + "learning_rate": 0.0006540432727022325, + "loss": 3.7833633422851562, + "step": 2540 + }, + { + "epoch": 35.29344978165939, + "grad_norm": 0.007855751551687717, + "learning_rate": 0.0006529575712776633, + "loss": 3.812349319458008, + "step": 2541 + }, + { + "epoch": 35.30742358078603, + "grad_norm": 0.007832377217710018, + "learning_rate": 0.0006518725643851091, + "loss": 3.7815146446228027, + "step": 2542 + }, + { + "epoch": 35.32139737991266, + "grad_norm": 0.006997724529355764, + "learning_rate": 0.0006507882528987395, + "loss": 3.795346975326538, + "step": 2543 + }, + { + "epoch": 35.3353711790393, + "grad_norm": 0.007433670572936535, + "learning_rate": 0.0006497046376921648, + "loss": 3.841017246246338, + "step": 2544 + }, + { + "epoch": 35.34934497816594, + "grad_norm": 0.0075767990201711655, + "learning_rate": 0.0006486217196384344, + "loss": 3.7913427352905273, + "step": 2545 + }, + { + "epoch": 35.36331877729258, + "grad_norm": 0.007551983464509249, + "learning_rate": 0.0006475394996100359, + "loss": 3.765105962753296, + "step": 2546 + }, + { + "epoch": 35.377292576419215, + "grad_norm": 0.007235546596348286, + "learning_rate": 0.0006464579784788943, + "loss": 3.7668676376342773, + "step": 2547 + }, + { + "epoch": 35.391266375545854, + "grad_norm": 0.007872921414673328, + "learning_rate": 0.0006453771571163716, + "loss": 3.770536422729492, + "step": 2548 + }, + { + "epoch": 35.40524017467249, + "grad_norm": 0.007598782889544964, + "learning_rate": 0.0006442970363932667, + "loss": 3.796818733215332, + "step": 2549 + }, + { + "epoch": 35.419213973799124, + "grad_norm": 0.007902889512479305, + "learning_rate": 0.0006432176171798124, + "loss": 3.824531078338623, + "step": 2550 + }, + { + "epoch": 35.43318777292576, + "grad_norm": 0.008964902721345425, + "learning_rate": 0.0006421389003456776, + "loss": 3.781778335571289, + "step": 2551 + }, + { + "epoch": 35.4471615720524, + "grad_norm": 0.009936625137925148, + "learning_rate": 0.000641060886759965, + "loss": 3.750248670578003, + "step": 2552 + }, + { + "epoch": 35.46113537117904, + "grad_norm": 0.008381214924156666, + "learning_rate": 0.0006399835772912106, + "loss": 3.797954797744751, + "step": 2553 + }, + { + "epoch": 35.47510917030568, + "grad_norm": 0.007752922363579273, + "learning_rate": 0.000638906972807384, + "loss": 3.789350748062134, + "step": 2554 + }, + { + "epoch": 35.48908296943232, + "grad_norm": 0.009501751512289047, + "learning_rate": 0.0006378310741758841, + "loss": 3.791079521179199, + "step": 2555 + }, + { + "epoch": 35.50305676855895, + "grad_norm": 0.00936005637049675, + "learning_rate": 0.0006367558822635453, + "loss": 3.78501033782959, + "step": 2556 + }, + { + "epoch": 35.51703056768559, + "grad_norm": 0.008370982483029366, + "learning_rate": 0.0006356813979366284, + "loss": 3.792647361755371, + "step": 2557 + }, + { + "epoch": 35.531004366812226, + "grad_norm": 0.00782831758260727, + "learning_rate": 0.000634607622060827, + "loss": 3.772569179534912, + "step": 2558 + }, + { + "epoch": 35.544978165938865, + "grad_norm": 0.00830658059567213, + "learning_rate": 0.0006335345555012627, + "loss": 3.7787656784057617, + "step": 2559 + }, + { + "epoch": 35.5589519650655, + "grad_norm": 0.009015833027660847, + "learning_rate": 0.000632462199122486, + "loss": 3.787418842315674, + "step": 2560 + }, + { + "epoch": 35.57292576419214, + "grad_norm": 0.008993860334157944, + "learning_rate": 0.0006313905537884752, + "loss": 3.818718433380127, + "step": 2561 + }, + { + "epoch": 35.58689956331878, + "grad_norm": 0.007954884320497513, + "learning_rate": 0.0006303196203626344, + "loss": 3.8100008964538574, + "step": 2562 + }, + { + "epoch": 35.60087336244541, + "grad_norm": 0.007742907386273146, + "learning_rate": 0.0006292493997077973, + "loss": 3.7786598205566406, + "step": 2563 + }, + { + "epoch": 35.61484716157205, + "grad_norm": 0.008748230524361134, + "learning_rate": 0.0006281798926862195, + "loss": 3.798635482788086, + "step": 2564 + }, + { + "epoch": 35.62882096069869, + "grad_norm": 0.008306438103318214, + "learning_rate": 0.000627111100159585, + "loss": 3.827213764190674, + "step": 2565 + }, + { + "epoch": 35.64279475982533, + "grad_norm": 0.008027350530028343, + "learning_rate": 0.0006260430229889987, + "loss": 3.7682337760925293, + "step": 2566 + }, + { + "epoch": 35.65676855895197, + "grad_norm": 0.007779776118695736, + "learning_rate": 0.0006249756620349931, + "loss": 3.7515475749969482, + "step": 2567 + }, + { + "epoch": 35.670742358078606, + "grad_norm": 0.008748390711843967, + "learning_rate": 0.000623909018157521, + "loss": 3.7963085174560547, + "step": 2568 + }, + { + "epoch": 35.68471615720524, + "grad_norm": 0.008587505668401718, + "learning_rate": 0.0006228430922159577, + "loss": 3.7767529487609863, + "step": 2569 + }, + { + "epoch": 35.698689956331876, + "grad_norm": 0.007172323763370514, + "learning_rate": 0.0006217778850691007, + "loss": 3.79768705368042, + "step": 2570 + }, + { + "epoch": 35.712663755458514, + "grad_norm": 0.007224083412438631, + "learning_rate": 0.000620713397575168, + "loss": 3.785390853881836, + "step": 2571 + }, + { + "epoch": 35.72663755458515, + "grad_norm": 0.008209837600588799, + "learning_rate": 0.0006196496305917987, + "loss": 3.809828281402588, + "step": 2572 + }, + { + "epoch": 35.74061135371179, + "grad_norm": 0.007760563865303993, + "learning_rate": 0.0006185865849760491, + "loss": 3.838716506958008, + "step": 2573 + }, + { + "epoch": 35.75458515283843, + "grad_norm": 0.007362260017544031, + "learning_rate": 0.0006175242615843977, + "loss": 3.7840137481689453, + "step": 2574 + }, + { + "epoch": 35.76855895196506, + "grad_norm": 0.008719335310161114, + "learning_rate": 0.0006164626612727381, + "loss": 3.8228893280029297, + "step": 2575 + }, + { + "epoch": 35.7825327510917, + "grad_norm": 0.00856029987335205, + "learning_rate": 0.0006154017848963826, + "loss": 3.80448317527771, + "step": 2576 + }, + { + "epoch": 35.79650655021834, + "grad_norm": 0.008078506216406822, + "learning_rate": 0.0006143416333100606, + "loss": 3.7881453037261963, + "step": 2577 + }, + { + "epoch": 35.81048034934498, + "grad_norm": 0.007957885973155499, + "learning_rate": 0.0006132822073679164, + "loss": 3.814870834350586, + "step": 2578 + }, + { + "epoch": 35.82445414847162, + "grad_norm": 0.009362339973449707, + "learning_rate": 0.0006122235079235114, + "loss": 3.798335075378418, + "step": 2579 + }, + { + "epoch": 35.838427947598255, + "grad_norm": 0.008272632025182247, + "learning_rate": 0.0006111655358298194, + "loss": 3.8212976455688477, + "step": 2580 + }, + { + "epoch": 35.852401746724894, + "grad_norm": 0.007548908703029156, + "learning_rate": 0.0006101082919392301, + "loss": 3.8312830924987793, + "step": 2581 + }, + { + "epoch": 35.866375545851525, + "grad_norm": 0.007543765474110842, + "learning_rate": 0.0006090517771035457, + "loss": 3.8302183151245117, + "step": 2582 + }, + { + "epoch": 35.880349344978164, + "grad_norm": 0.007520351093262434, + "learning_rate": 0.000607995992173981, + "loss": 3.7898802757263184, + "step": 2583 + }, + { + "epoch": 35.8943231441048, + "grad_norm": 0.007168850861489773, + "learning_rate": 0.000606940938001163, + "loss": 3.8078577518463135, + "step": 2584 + }, + { + "epoch": 35.90829694323144, + "grad_norm": 0.007420001085847616, + "learning_rate": 0.0006058866154351296, + "loss": 3.799571990966797, + "step": 2585 + }, + { + "epoch": 35.92227074235808, + "grad_norm": 0.007621012162417173, + "learning_rate": 0.0006048330253253303, + "loss": 3.833843946456909, + "step": 2586 + }, + { + "epoch": 35.93624454148472, + "grad_norm": 0.007661828305572271, + "learning_rate": 0.0006037801685206222, + "loss": 3.820556640625, + "step": 2587 + }, + { + "epoch": 35.95021834061135, + "grad_norm": 0.007657836657017469, + "learning_rate": 0.0006027280458692738, + "loss": 3.798994541168213, + "step": 2588 + }, + { + "epoch": 35.96419213973799, + "grad_norm": 0.007722760550677776, + "learning_rate": 0.0006016766582189613, + "loss": 3.8589251041412354, + "step": 2589 + }, + { + "epoch": 35.97816593886463, + "grad_norm": 0.00789247453212738, + "learning_rate": 0.0006006260064167685, + "loss": 3.8673787117004395, + "step": 2590 + }, + { + "epoch": 35.992139737991266, + "grad_norm": 0.009137817658483982, + "learning_rate": 0.0005995760913091874, + "loss": 3.801938056945801, + "step": 2591 + }, + { + "epoch": 36.0, + "grad_norm": 0.008374894969165325, + "learning_rate": 0.0005985269137421141, + "loss": 3.820045232772827, + "step": 2592 + }, + { + "epoch": 36.0, + "eval_loss": 4.3730878829956055, + "eval_runtime": 48.1607, + "eval_samples_per_second": 50.705, + "eval_steps_per_second": 1.599, + "step": 2592 + }, + { + "epoch": 36.01397379912664, + "grad_norm": 0.009492472745478153, + "learning_rate": 0.0005974784745608536, + "loss": 3.738617420196533, + "step": 2593 + }, + { + "epoch": 36.02794759825328, + "grad_norm": 0.010017318651080132, + "learning_rate": 0.0005964307746101135, + "loss": 3.7551231384277344, + "step": 2594 + }, + { + "epoch": 36.041921397379916, + "grad_norm": 0.008958659134805202, + "learning_rate": 0.000595383814734007, + "loss": 3.7739763259887695, + "step": 2595 + }, + { + "epoch": 36.05589519650655, + "grad_norm": 0.008034365251660347, + "learning_rate": 0.0005943375957760505, + "loss": 3.7678627967834473, + "step": 2596 + }, + { + "epoch": 36.069868995633186, + "grad_norm": 0.008489780128002167, + "learning_rate": 0.000593292118579164, + "loss": 3.720778465270996, + "step": 2597 + }, + { + "epoch": 36.083842794759825, + "grad_norm": 0.009107839316129684, + "learning_rate": 0.0005922473839856699, + "loss": 3.7224440574645996, + "step": 2598 + }, + { + "epoch": 36.09781659388646, + "grad_norm": 0.009295578114688396, + "learning_rate": 0.0005912033928372906, + "loss": 3.7695257663726807, + "step": 2599 + }, + { + "epoch": 36.1117903930131, + "grad_norm": 0.007896862924098969, + "learning_rate": 0.0005901601459751532, + "loss": 3.742567300796509, + "step": 2600 + }, + { + "epoch": 36.12576419213974, + "grad_norm": 0.008227395825088024, + "learning_rate": 0.0005891176442397808, + "loss": 3.7547202110290527, + "step": 2601 + }, + { + "epoch": 36.13973799126637, + "grad_norm": 0.009326796047389507, + "learning_rate": 0.0005880758884710992, + "loss": 3.724808931350708, + "step": 2602 + }, + { + "epoch": 36.15371179039301, + "grad_norm": 0.007786506786942482, + "learning_rate": 0.0005870348795084319, + "loss": 3.7403640747070312, + "step": 2603 + }, + { + "epoch": 36.16768558951965, + "grad_norm": 0.00821041688323021, + "learning_rate": 0.0005859946181905009, + "loss": 3.770012617111206, + "step": 2604 + }, + { + "epoch": 36.18165938864629, + "grad_norm": 0.009544595144689083, + "learning_rate": 0.000584955105355427, + "loss": 3.7378525733947754, + "step": 2605 + }, + { + "epoch": 36.19563318777293, + "grad_norm": 0.008973932825028896, + "learning_rate": 0.0005839163418407254, + "loss": 3.7184295654296875, + "step": 2606 + }, + { + "epoch": 36.209606986899566, + "grad_norm": 0.009023688733577728, + "learning_rate": 0.00058287832848331, + "loss": 3.7388229370117188, + "step": 2607 + }, + { + "epoch": 36.223580786026204, + "grad_norm": 0.008151066489517689, + "learning_rate": 0.0005818410661194889, + "loss": 3.7632875442504883, + "step": 2608 + }, + { + "epoch": 36.237554585152836, + "grad_norm": 0.008495030924677849, + "learning_rate": 0.0005808045555849668, + "loss": 3.8114736080169678, + "step": 2609 + }, + { + "epoch": 36.251528384279474, + "grad_norm": 0.008739210665225983, + "learning_rate": 0.0005797687977148394, + "loss": 3.7647829055786133, + "step": 2610 + }, + { + "epoch": 36.26550218340611, + "grad_norm": 0.008062886074185371, + "learning_rate": 0.0005787337933435999, + "loss": 3.782172441482544, + "step": 2611 + }, + { + "epoch": 36.27947598253275, + "grad_norm": 0.007615398149937391, + "learning_rate": 0.0005776995433051326, + "loss": 3.7564642429351807, + "step": 2612 + }, + { + "epoch": 36.29344978165939, + "grad_norm": 0.007938290014863014, + "learning_rate": 0.0005766660484327133, + "loss": 3.756208896636963, + "step": 2613 + }, + { + "epoch": 36.30742358078603, + "grad_norm": 0.008517666719853878, + "learning_rate": 0.0005756333095590102, + "loss": 3.76729416847229, + "step": 2614 + }, + { + "epoch": 36.32139737991266, + "grad_norm": 0.007426336407661438, + "learning_rate": 0.0005746013275160827, + "loss": 3.74001407623291, + "step": 2615 + }, + { + "epoch": 36.3353711790393, + "grad_norm": 0.007664753124117851, + "learning_rate": 0.0005735701031353804, + "loss": 3.7819876670837402, + "step": 2616 + }, + { + "epoch": 36.34934497816594, + "grad_norm": 0.007903067395091057, + "learning_rate": 0.0005725396372477414, + "loss": 3.781219959259033, + "step": 2617 + }, + { + "epoch": 36.36331877729258, + "grad_norm": 0.007251975126564503, + "learning_rate": 0.0005715099306833931, + "loss": 3.763620615005493, + "step": 2618 + }, + { + "epoch": 36.377292576419215, + "grad_norm": 0.007422176189720631, + "learning_rate": 0.0005704809842719533, + "loss": 3.775099754333496, + "step": 2619 + }, + { + "epoch": 36.391266375545854, + "grad_norm": 0.009161600843071938, + "learning_rate": 0.0005694527988424239, + "loss": 3.792268991470337, + "step": 2620 + }, + { + "epoch": 36.40524017467249, + "grad_norm": 0.008330529555678368, + "learning_rate": 0.0005684253752231958, + "loss": 3.7967824935913086, + "step": 2621 + }, + { + "epoch": 36.419213973799124, + "grad_norm": 0.0073860278353095055, + "learning_rate": 0.0005673987142420458, + "loss": 3.76505970954895, + "step": 2622 + }, + { + "epoch": 36.43318777292576, + "grad_norm": 0.008462096564471722, + "learning_rate": 0.0005663728167261363, + "loss": 3.740450143814087, + "step": 2623 + }, + { + "epoch": 36.4471615720524, + "grad_norm": 0.00838128849864006, + "learning_rate": 0.0005653476835020138, + "loss": 3.7863235473632812, + "step": 2624 + }, + { + "epoch": 36.46113537117904, + "grad_norm": 0.007381014991551638, + "learning_rate": 0.0005643233153956092, + "loss": 3.772179126739502, + "step": 2625 + }, + { + "epoch": 36.47510917030568, + "grad_norm": 0.007363761775195599, + "learning_rate": 0.0005632997132322392, + "loss": 3.7793867588043213, + "step": 2626 + }, + { + "epoch": 36.48908296943232, + "grad_norm": 0.008684969507157803, + "learning_rate": 0.0005622768778366001, + "loss": 3.7736339569091797, + "step": 2627 + }, + { + "epoch": 36.50305676855895, + "grad_norm": 0.007934059016406536, + "learning_rate": 0.0005612548100327722, + "loss": 3.8040053844451904, + "step": 2628 + }, + { + "epoch": 36.51703056768559, + "grad_norm": 0.007136524189263582, + "learning_rate": 0.0005602335106442172, + "loss": 3.7589468955993652, + "step": 2629 + }, + { + "epoch": 36.531004366812226, + "grad_norm": 0.00789916142821312, + "learning_rate": 0.0005592129804937779, + "loss": 3.7953195571899414, + "step": 2630 + }, + { + "epoch": 36.544978165938865, + "grad_norm": 0.007919179275631905, + "learning_rate": 0.0005581932204036761, + "loss": 3.7539429664611816, + "step": 2631 + }, + { + "epoch": 36.5589519650655, + "grad_norm": 0.007734223268926144, + "learning_rate": 0.0005571742311955147, + "loss": 3.7620325088500977, + "step": 2632 + }, + { + "epoch": 36.57292576419214, + "grad_norm": 0.007005508057773113, + "learning_rate": 0.0005561560136902749, + "loss": 3.762538433074951, + "step": 2633 + }, + { + "epoch": 36.58689956331878, + "grad_norm": 0.008137146942317486, + "learning_rate": 0.0005551385687083161, + "loss": 3.773789405822754, + "step": 2634 + }, + { + "epoch": 36.60087336244541, + "grad_norm": 0.007310070563107729, + "learning_rate": 0.0005541218970693757, + "loss": 3.8077168464660645, + "step": 2635 + }, + { + "epoch": 36.61484716157205, + "grad_norm": 0.007335536181926727, + "learning_rate": 0.0005531059995925661, + "loss": 3.7761921882629395, + "step": 2636 + }, + { + "epoch": 36.62882096069869, + "grad_norm": 0.008022893220186234, + "learning_rate": 0.00055209087709638, + "loss": 3.7769880294799805, + "step": 2637 + }, + { + "epoch": 36.64279475982533, + "grad_norm": 0.008027128875255585, + "learning_rate": 0.0005510765303986814, + "loss": 3.7709951400756836, + "step": 2638 + }, + { + "epoch": 36.65676855895197, + "grad_norm": 0.00697955721989274, + "learning_rate": 0.0005500629603167118, + "loss": 3.7855210304260254, + "step": 2639 + }, + { + "epoch": 36.670742358078606, + "grad_norm": 0.007455077022314072, + "learning_rate": 0.0005490501676670861, + "loss": 3.7850632667541504, + "step": 2640 + }, + { + "epoch": 36.68471615720524, + "grad_norm": 0.0069657545536756516, + "learning_rate": 0.0005480381532657932, + "loss": 3.742671251296997, + "step": 2641 + }, + { + "epoch": 36.698689956331876, + "grad_norm": 0.007609451189637184, + "learning_rate": 0.0005470269179281953, + "loss": 3.793771266937256, + "step": 2642 + }, + { + "epoch": 36.712663755458514, + "grad_norm": 0.007437967229634523, + "learning_rate": 0.0005460164624690255, + "loss": 3.778846263885498, + "step": 2643 + }, + { + "epoch": 36.72663755458515, + "grad_norm": 0.007123028859496117, + "learning_rate": 0.00054500678770239, + "loss": 3.768375873565674, + "step": 2644 + }, + { + "epoch": 36.74061135371179, + "grad_norm": 0.008176218718290329, + "learning_rate": 0.0005439978944417654, + "loss": 3.82069730758667, + "step": 2645 + }, + { + "epoch": 36.75458515283843, + "grad_norm": 0.006943001411855221, + "learning_rate": 0.000542989783499999, + "loss": 3.7731552124023438, + "step": 2646 + }, + { + "epoch": 36.76855895196506, + "grad_norm": 0.0072205523028969765, + "learning_rate": 0.0005419824556893069, + "loss": 3.7724058628082275, + "step": 2647 + }, + { + "epoch": 36.7825327510917, + "grad_norm": 0.007656400091946125, + "learning_rate": 0.0005409759118212758, + "loss": 3.780989646911621, + "step": 2648 + }, + { + "epoch": 36.79650655021834, + "grad_norm": 0.006972154602408409, + "learning_rate": 0.0005399701527068599, + "loss": 3.7971787452697754, + "step": 2649 + }, + { + "epoch": 36.81048034934498, + "grad_norm": 0.007674331311136484, + "learning_rate": 0.0005389651791563806, + "loss": 3.798673391342163, + "step": 2650 + }, + { + "epoch": 36.82445414847162, + "grad_norm": 0.006952045019716024, + "learning_rate": 0.0005379609919795268, + "loss": 3.7982821464538574, + "step": 2651 + }, + { + "epoch": 36.838427947598255, + "grad_norm": 0.007882075384259224, + "learning_rate": 0.0005369575919853545, + "loss": 3.825962781906128, + "step": 2652 + }, + { + "epoch": 36.852401746724894, + "grad_norm": 0.007143386639654636, + "learning_rate": 0.0005359549799822854, + "loss": 3.826598644256592, + "step": 2653 + }, + { + "epoch": 36.866375545851525, + "grad_norm": 0.007587207015603781, + "learning_rate": 0.0005349531567781041, + "loss": 3.80623197555542, + "step": 2654 + }, + { + "epoch": 36.880349344978164, + "grad_norm": 0.007677781395614147, + "learning_rate": 0.0005339521231799632, + "loss": 3.7987256050109863, + "step": 2655 + }, + { + "epoch": 36.8943231441048, + "grad_norm": 0.007372343447059393, + "learning_rate": 0.0005329518799943773, + "loss": 3.7733969688415527, + "step": 2656 + }, + { + "epoch": 36.90829694323144, + "grad_norm": 0.00861304346472025, + "learning_rate": 0.0005319524280272231, + "loss": 3.7903685569763184, + "step": 2657 + }, + { + "epoch": 36.92227074235808, + "grad_norm": 0.00816237274557352, + "learning_rate": 0.0005309537680837418, + "loss": 3.77058744430542, + "step": 2658 + }, + { + "epoch": 36.93624454148472, + "grad_norm": 0.007624274119734764, + "learning_rate": 0.0005299559009685356, + "loss": 3.787982940673828, + "step": 2659 + }, + { + "epoch": 36.95021834061135, + "grad_norm": 0.007416521664708853, + "learning_rate": 0.0005289588274855683, + "loss": 3.7991762161254883, + "step": 2660 + }, + { + "epoch": 36.96419213973799, + "grad_norm": 0.007461614441126585, + "learning_rate": 0.0005279625484381632, + "loss": 3.8036863803863525, + "step": 2661 + }, + { + "epoch": 36.97816593886463, + "grad_norm": 0.007766335736960173, + "learning_rate": 0.0005269670646290043, + "loss": 3.809243679046631, + "step": 2662 + }, + { + "epoch": 36.992139737991266, + "grad_norm": 0.0073394919745624065, + "learning_rate": 0.0005259723768601365, + "loss": 3.756606101989746, + "step": 2663 + }, + { + "epoch": 37.0, + "grad_norm": 0.008771199733018875, + "learning_rate": 0.0005249784859329602, + "loss": 3.773951768875122, + "step": 2664 + }, + { + "epoch": 37.0, + "eval_loss": 4.370637893676758, + "eval_runtime": 49.4904, + "eval_samples_per_second": 49.343, + "eval_steps_per_second": 1.556, + "step": 2664 + }, + { + "epoch": 37.01397379912664, + "grad_norm": 0.008164486847817898, + "learning_rate": 0.0005239853926482358, + "loss": 3.7619595527648926, + "step": 2665 + }, + { + "epoch": 37.02794759825328, + "grad_norm": 0.007980132475495338, + "learning_rate": 0.0005229930978060806, + "loss": 3.7398204803466797, + "step": 2666 + }, + { + "epoch": 37.041921397379916, + "grad_norm": 0.0077056242153048515, + "learning_rate": 0.0005220016022059692, + "loss": 3.699580669403076, + "step": 2667 + }, + { + "epoch": 37.05589519650655, + "grad_norm": 0.0074698119424283504, + "learning_rate": 0.0005210109066467308, + "loss": 3.7299928665161133, + "step": 2668 + }, + { + "epoch": 37.069868995633186, + "grad_norm": 0.008384468965232372, + "learning_rate": 0.0005200210119265512, + "loss": 3.704989433288574, + "step": 2669 + }, + { + "epoch": 37.083842794759825, + "grad_norm": 0.008066141977906227, + "learning_rate": 0.0005190319188429709, + "loss": 3.751697540283203, + "step": 2670 + }, + { + "epoch": 37.09781659388646, + "grad_norm": 0.007213297300040722, + "learning_rate": 0.0005180436281928841, + "loss": 3.740692377090454, + "step": 2671 + }, + { + "epoch": 37.1117903930131, + "grad_norm": 0.007343464996665716, + "learning_rate": 0.0005170561407725396, + "loss": 3.7060189247131348, + "step": 2672 + }, + { + "epoch": 37.12576419213974, + "grad_norm": 0.00783725269138813, + "learning_rate": 0.0005160694573775363, + "loss": 3.7414426803588867, + "step": 2673 + }, + { + "epoch": 37.13973799126637, + "grad_norm": 0.007640232797712088, + "learning_rate": 0.0005150835788028294, + "loss": 3.7451486587524414, + "step": 2674 + }, + { + "epoch": 37.15371179039301, + "grad_norm": 0.008129071444272995, + "learning_rate": 0.0005140985058427218, + "loss": 3.734839916229248, + "step": 2675 + }, + { + "epoch": 37.16768558951965, + "grad_norm": 0.006920623127371073, + "learning_rate": 0.0005131142392908697, + "loss": 3.7138614654541016, + "step": 2676 + }, + { + "epoch": 37.18165938864629, + "grad_norm": 0.00798740889877081, + "learning_rate": 0.0005121307799402787, + "loss": 3.76496958732605, + "step": 2677 + }, + { + "epoch": 37.19563318777293, + "grad_norm": 0.006689135450869799, + "learning_rate": 0.0005111481285833045, + "loss": 3.7707948684692383, + "step": 2678 + }, + { + "epoch": 37.209606986899566, + "grad_norm": 0.007544311694800854, + "learning_rate": 0.0005101662860116516, + "loss": 3.7393479347229004, + "step": 2679 + }, + { + "epoch": 37.223580786026204, + "grad_norm": 0.007487910334020853, + "learning_rate": 0.0005091852530163717, + "loss": 3.7486305236816406, + "step": 2680 + }, + { + "epoch": 37.237554585152836, + "grad_norm": 0.00772349676117301, + "learning_rate": 0.0005082050303878673, + "loss": 3.753629207611084, + "step": 2681 + }, + { + "epoch": 37.251528384279474, + "grad_norm": 0.007883663289248943, + "learning_rate": 0.0005072256189158844, + "loss": 3.7497611045837402, + "step": 2682 + }, + { + "epoch": 37.26550218340611, + "grad_norm": 0.008153906092047691, + "learning_rate": 0.0005062470193895178, + "loss": 3.7459521293640137, + "step": 2683 + }, + { + "epoch": 37.27947598253275, + "grad_norm": 0.007083641365170479, + "learning_rate": 0.0005052692325972075, + "loss": 3.725599765777588, + "step": 2684 + }, + { + "epoch": 37.29344978165939, + "grad_norm": 0.007102683652192354, + "learning_rate": 0.0005042922593267385, + "loss": 3.7724428176879883, + "step": 2685 + }, + { + "epoch": 37.30742358078603, + "grad_norm": 0.008104706183075905, + "learning_rate": 0.0005033161003652409, + "loss": 3.7386724948883057, + "step": 2686 + }, + { + "epoch": 37.32139737991266, + "grad_norm": 0.0073765418492257595, + "learning_rate": 0.0005023407564991878, + "loss": 3.7478132247924805, + "step": 2687 + }, + { + "epoch": 37.3353711790393, + "grad_norm": 0.007110628765076399, + "learning_rate": 0.0005013662285143963, + "loss": 3.771925687789917, + "step": 2688 + }, + { + "epoch": 37.34934497816594, + "grad_norm": 0.007323848083615303, + "learning_rate": 0.000500392517196026, + "loss": 3.7460904121398926, + "step": 2689 + }, + { + "epoch": 37.36331877729258, + "grad_norm": 0.007830807007849216, + "learning_rate": 0.0004994196233285786, + "loss": 3.751222610473633, + "step": 2690 + }, + { + "epoch": 37.377292576419215, + "grad_norm": 0.007710007485002279, + "learning_rate": 0.0004984475476958973, + "loss": 3.739858627319336, + "step": 2691 + }, + { + "epoch": 37.391266375545854, + "grad_norm": 0.007293973118066788, + "learning_rate": 0.0004974762910811655, + "loss": 3.7410120964050293, + "step": 2692 + }, + { + "epoch": 37.40524017467249, + "grad_norm": 0.0073824538849294186, + "learning_rate": 0.0004965058542669078, + "loss": 3.711397647857666, + "step": 2693 + }, + { + "epoch": 37.419213973799124, + "grad_norm": 0.007156423293054104, + "learning_rate": 0.0004955362380349866, + "loss": 3.7881593704223633, + "step": 2694 + }, + { + "epoch": 37.43318777292576, + "grad_norm": 0.007705818396061659, + "learning_rate": 0.0004945674431666047, + "loss": 3.7639269828796387, + "step": 2695 + }, + { + "epoch": 37.4471615720524, + "grad_norm": 0.0075300210155546665, + "learning_rate": 0.0004935994704423026, + "loss": 3.7562999725341797, + "step": 2696 + }, + { + "epoch": 37.46113537117904, + "grad_norm": 0.007137116510421038, + "learning_rate": 0.000492632320641959, + "loss": 3.7659149169921875, + "step": 2697 + }, + { + "epoch": 37.47510917030568, + "grad_norm": 0.007225630804896355, + "learning_rate": 0.0004916659945447878, + "loss": 3.7625155448913574, + "step": 2698 + }, + { + "epoch": 37.48908296943232, + "grad_norm": 0.008027992211282253, + "learning_rate": 0.0004907004929293408, + "loss": 3.746478796005249, + "step": 2699 + }, + { + "epoch": 37.50305676855895, + "grad_norm": 0.007582477759569883, + "learning_rate": 0.0004897358165735066, + "loss": 3.759408950805664, + "step": 2700 + }, + { + "epoch": 37.51703056768559, + "grad_norm": 0.007575645577162504, + "learning_rate": 0.000488771966254506, + "loss": 3.7430644035339355, + "step": 2701 + }, + { + "epoch": 37.531004366812226, + "grad_norm": 0.008404700085520744, + "learning_rate": 0.0004878089427488963, + "loss": 3.7727198600769043, + "step": 2702 + }, + { + "epoch": 37.544978165938865, + "grad_norm": 0.007118664216250181, + "learning_rate": 0.00048684674683256815, + "loss": 3.747460126876831, + "step": 2703 + }, + { + "epoch": 37.5589519650655, + "grad_norm": 0.007220778148621321, + "learning_rate": 0.00048588537928074584, + "loss": 3.746413230895996, + "step": 2704 + }, + { + "epoch": 37.57292576419214, + "grad_norm": 0.0076194945722818375, + "learning_rate": 0.0004849248408679849, + "loss": 3.7808125019073486, + "step": 2705 + }, + { + "epoch": 37.58689956331878, + "grad_norm": 0.007285538129508495, + "learning_rate": 0.00048396513236817387, + "loss": 3.7750301361083984, + "step": 2706 + }, + { + "epoch": 37.60087336244541, + "grad_norm": 0.007698678877204657, + "learning_rate": 0.00048300625455453383, + "loss": 3.7474594116210938, + "step": 2707 + }, + { + "epoch": 37.61484716157205, + "grad_norm": 0.007687424309551716, + "learning_rate": 0.0004820482081996141, + "loss": 3.7929720878601074, + "step": 2708 + }, + { + "epoch": 37.62882096069869, + "grad_norm": 0.007363605313003063, + "learning_rate": 0.0004810909940752959, + "loss": 3.75911021232605, + "step": 2709 + }, + { + "epoch": 37.64279475982533, + "grad_norm": 0.007791243493556976, + "learning_rate": 0.00048013461295278836, + "loss": 3.7567291259765625, + "step": 2710 + }, + { + "epoch": 37.65676855895197, + "grad_norm": 0.007375436834990978, + "learning_rate": 0.0004791790656026323, + "loss": 3.7784414291381836, + "step": 2711 + }, + { + "epoch": 37.670742358078606, + "grad_norm": 0.007330593653023243, + "learning_rate": 0.0004782243527946934, + "loss": 3.7533810138702393, + "step": 2712 + }, + { + "epoch": 37.68471615720524, + "grad_norm": 0.006926248781383038, + "learning_rate": 0.00047727047529816694, + "loss": 3.7795910835266113, + "step": 2713 + }, + { + "epoch": 37.698689956331876, + "grad_norm": 0.007102136500179768, + "learning_rate": 0.00047631743388157496, + "loss": 3.812549114227295, + "step": 2714 + }, + { + "epoch": 37.712663755458514, + "grad_norm": 0.006676106248050928, + "learning_rate": 0.00047536522931276555, + "loss": 3.7578792572021484, + "step": 2715 + }, + { + "epoch": 37.72663755458515, + "grad_norm": 0.006926767062395811, + "learning_rate": 0.0004744138623589134, + "loss": 3.782416820526123, + "step": 2716 + }, + { + "epoch": 37.74061135371179, + "grad_norm": 0.006894777063280344, + "learning_rate": 0.000473463333786516, + "loss": 3.7840042114257812, + "step": 2717 + }, + { + "epoch": 37.75458515283843, + "grad_norm": 0.007393381092697382, + "learning_rate": 0.00047251364436139923, + "loss": 3.7532780170440674, + "step": 2718 + }, + { + "epoch": 37.76855895196506, + "grad_norm": 0.006510896608233452, + "learning_rate": 0.0004715647948487092, + "loss": 3.7618460655212402, + "step": 2719 + }, + { + "epoch": 37.7825327510917, + "grad_norm": 0.007283828221261501, + "learning_rate": 0.0004706167860129171, + "loss": 3.768498420715332, + "step": 2720 + }, + { + "epoch": 37.79650655021834, + "grad_norm": 0.00655644154176116, + "learning_rate": 0.0004696696186178168, + "loss": 3.7724053859710693, + "step": 2721 + }, + { + "epoch": 37.81048034934498, + "grad_norm": 0.007338619790971279, + "learning_rate": 0.0004687232934265239, + "loss": 3.7814595699310303, + "step": 2722 + }, + { + "epoch": 37.82445414847162, + "grad_norm": 0.006839970126748085, + "learning_rate": 0.0004677778112014758, + "loss": 3.771808624267578, + "step": 2723 + }, + { + "epoch": 37.838427947598255, + "grad_norm": 0.006776005961000919, + "learning_rate": 0.0004668331727044297, + "loss": 3.786924123764038, + "step": 2724 + }, + { + "epoch": 37.852401746724894, + "grad_norm": 0.007011357229202986, + "learning_rate": 0.00046588937869646415, + "loss": 3.787672996520996, + "step": 2725 + }, + { + "epoch": 37.866375545851525, + "grad_norm": 0.006482086144387722, + "learning_rate": 0.0004649464299379771, + "loss": 3.7794556617736816, + "step": 2726 + }, + { + "epoch": 37.880349344978164, + "grad_norm": 0.0069654518738389015, + "learning_rate": 0.000464004327188685, + "loss": 3.7520360946655273, + "step": 2727 + }, + { + "epoch": 37.8943231441048, + "grad_norm": 0.006526473443955183, + "learning_rate": 0.0004630630712076235, + "loss": 3.734144926071167, + "step": 2728 + }, + { + "epoch": 37.90829694323144, + "grad_norm": 0.006822322495281696, + "learning_rate": 0.00046212266275314504, + "loss": 3.7614707946777344, + "step": 2729 + }, + { + "epoch": 37.92227074235808, + "grad_norm": 0.006827171426266432, + "learning_rate": 0.00046118310258292045, + "loss": 3.755218029022217, + "step": 2730 + }, + { + "epoch": 37.93624454148472, + "grad_norm": 0.00706718722358346, + "learning_rate": 0.00046024439145393546, + "loss": 3.7397823333740234, + "step": 2731 + }, + { + "epoch": 37.95021834061135, + "grad_norm": 0.00689381267875433, + "learning_rate": 0.00045930653012249314, + "loss": 3.758922576904297, + "step": 2732 + }, + { + "epoch": 37.96419213973799, + "grad_norm": 0.007069882471114397, + "learning_rate": 0.0004583695193442113, + "loss": 3.759767532348633, + "step": 2733 + }, + { + "epoch": 37.97816593886463, + "grad_norm": 0.006889182608574629, + "learning_rate": 0.0004574333598740228, + "loss": 3.7555460929870605, + "step": 2734 + }, + { + "epoch": 37.992139737991266, + "grad_norm": 0.006866258103400469, + "learning_rate": 0.00045649805246617475, + "loss": 3.762068271636963, + "step": 2735 + }, + { + "epoch": 38.0, + "grad_norm": 0.008696220815181732, + "learning_rate": 0.000455563597874226, + "loss": 3.7110466957092285, + "step": 2736 + }, + { + "epoch": 38.0, + "eval_loss": 4.369833469390869, + "eval_runtime": 48.7852, + "eval_samples_per_second": 50.056, + "eval_steps_per_second": 1.578, + "step": 2736 + }, + { + "epoch": 38.01397379912664, + "grad_norm": 0.0077184769324958324, + "learning_rate": 0.0004546299968510516, + "loss": 3.704669713973999, + "step": 2737 + }, + { + "epoch": 38.02794759825328, + "grad_norm": 0.00846019759774208, + "learning_rate": 0.0004536972501488358, + "loss": 3.701132297515869, + "step": 2738 + }, + { + "epoch": 38.041921397379916, + "grad_norm": 0.008113945834338665, + "learning_rate": 0.00045276535851907613, + "loss": 3.699582099914551, + "step": 2739 + }, + { + "epoch": 38.05589519650655, + "grad_norm": 0.0072370911948382854, + "learning_rate": 0.0004518343227125807, + "loss": 3.698126792907715, + "step": 2740 + }, + { + "epoch": 38.069868995633186, + "grad_norm": 0.00817187875509262, + "learning_rate": 0.0004509041434794685, + "loss": 3.738393783569336, + "step": 2741 + }, + { + "epoch": 38.083842794759825, + "grad_norm": 0.007495569530874491, + "learning_rate": 0.00044997482156916845, + "loss": 3.727083683013916, + "step": 2742 + }, + { + "epoch": 38.09781659388646, + "grad_norm": 0.007855897769331932, + "learning_rate": 0.0004490463577304171, + "loss": 3.716137170791626, + "step": 2743 + }, + { + "epoch": 38.1117903930131, + "grad_norm": 0.007817867211997509, + "learning_rate": 0.0004481187527112628, + "loss": 3.750791072845459, + "step": 2744 + }, + { + "epoch": 38.12576419213974, + "grad_norm": 0.007712380029261112, + "learning_rate": 0.00044719200725905887, + "loss": 3.720554828643799, + "step": 2745 + }, + { + "epoch": 38.13973799126637, + "grad_norm": 0.00859921332448721, + "learning_rate": 0.00044626612212046816, + "loss": 3.742496967315674, + "step": 2746 + }, + { + "epoch": 38.15371179039301, + "grad_norm": 0.007692081853747368, + "learning_rate": 0.0004453410980414582, + "loss": 3.703803300857544, + "step": 2747 + }, + { + "epoch": 38.16768558951965, + "grad_norm": 0.007734883576631546, + "learning_rate": 0.00044441693576730565, + "loss": 3.7371301651000977, + "step": 2748 + }, + { + "epoch": 38.18165938864629, + "grad_norm": 0.009022004902362823, + "learning_rate": 0.0004434936360425916, + "loss": 3.727294683456421, + "step": 2749 + }, + { + "epoch": 38.19563318777293, + "grad_norm": 0.008058780804276466, + "learning_rate": 0.00044257119961120067, + "loss": 3.7583885192871094, + "step": 2750 + }, + { + "epoch": 38.209606986899566, + "grad_norm": 0.00698013324290514, + "learning_rate": 0.00044164962721632376, + "loss": 3.7312440872192383, + "step": 2751 + }, + { + "epoch": 38.223580786026204, + "grad_norm": 0.007999536581337452, + "learning_rate": 0.00044072891960045523, + "loss": 3.6972708702087402, + "step": 2752 + }, + { + "epoch": 38.237554585152836, + "grad_norm": 0.008125107735395432, + "learning_rate": 0.0004398090775053929, + "loss": 3.7123398780822754, + "step": 2753 + }, + { + "epoch": 38.251528384279474, + "grad_norm": 0.006736223120242357, + "learning_rate": 0.0004388901016722357, + "loss": 3.749678134918213, + "step": 2754 + }, + { + "epoch": 38.26550218340611, + "grad_norm": 0.009154478088021278, + "learning_rate": 0.000437971992841387, + "loss": 3.722728967666626, + "step": 2755 + }, + { + "epoch": 38.27947598253275, + "grad_norm": 0.008070871233940125, + "learning_rate": 0.00043705475175255087, + "loss": 3.7297017574310303, + "step": 2756 + }, + { + "epoch": 38.29344978165939, + "grad_norm": 0.007498566992580891, + "learning_rate": 0.0004361383791447307, + "loss": 3.7286131381988525, + "step": 2757 + }, + { + "epoch": 38.30742358078603, + "grad_norm": 0.007642887998372316, + "learning_rate": 0.00043522287575623183, + "loss": 3.750898838043213, + "step": 2758 + }, + { + "epoch": 38.32139737991266, + "grad_norm": 0.007800443097949028, + "learning_rate": 0.0004343082423246588, + "loss": 3.719512462615967, + "step": 2759 + }, + { + "epoch": 38.3353711790393, + "grad_norm": 0.007393732666969299, + "learning_rate": 0.0004333944795869159, + "loss": 3.7136006355285645, + "step": 2760 + }, + { + "epoch": 38.34934497816594, + "grad_norm": 0.008559555746614933, + "learning_rate": 0.0004324815882792042, + "loss": 3.7087020874023438, + "step": 2761 + }, + { + "epoch": 38.36331877729258, + "grad_norm": 0.007379850372672081, + "learning_rate": 0.0004315695691370241, + "loss": 3.7638165950775146, + "step": 2762 + }, + { + "epoch": 38.377292576419215, + "grad_norm": 0.007374506443738937, + "learning_rate": 0.00043065842289517303, + "loss": 3.7208821773529053, + "step": 2763 + }, + { + "epoch": 38.391266375545854, + "grad_norm": 0.00755207147449255, + "learning_rate": 0.0004297481502877451, + "loss": 3.738490581512451, + "step": 2764 + }, + { + "epoch": 38.40524017467249, + "grad_norm": 0.007752610836178064, + "learning_rate": 0.0004288387520481306, + "loss": 3.7401833534240723, + "step": 2765 + }, + { + "epoch": 38.419213973799124, + "grad_norm": 0.006907653529196978, + "learning_rate": 0.0004279302289090151, + "loss": 3.7754998207092285, + "step": 2766 + }, + { + "epoch": 38.43318777292576, + "grad_norm": 0.007771312724798918, + "learning_rate": 0.0004270225816023797, + "loss": 3.73626446723938, + "step": 2767 + }, + { + "epoch": 38.4471615720524, + "grad_norm": 0.007395982276648283, + "learning_rate": 0.0004261158108594987, + "loss": 3.7159175872802734, + "step": 2768 + }, + { + "epoch": 38.46113537117904, + "grad_norm": 0.007196040358394384, + "learning_rate": 0.0004252099174109411, + "loss": 3.750246047973633, + "step": 2769 + }, + { + "epoch": 38.47510917030568, + "grad_norm": 0.008241173811256886, + "learning_rate": 0.0004243049019865692, + "loss": 3.70859694480896, + "step": 2770 + }, + { + "epoch": 38.48908296943232, + "grad_norm": 0.0072947596199810505, + "learning_rate": 0.00042340076531553755, + "loss": 3.7385268211364746, + "step": 2771 + }, + { + "epoch": 38.50305676855895, + "grad_norm": 0.00782464537769556, + "learning_rate": 0.00042249750812629297, + "loss": 3.75885272026062, + "step": 2772 + }, + { + "epoch": 38.51703056768559, + "grad_norm": 0.007813232019543648, + "learning_rate": 0.0004215951311465722, + "loss": 3.71810245513916, + "step": 2773 + }, + { + "epoch": 38.531004366812226, + "grad_norm": 0.006900500506162643, + "learning_rate": 0.00042069363510340604, + "loss": 3.7365174293518066, + "step": 2774 + }, + { + "epoch": 38.544978165938865, + "grad_norm": 0.007812316995114088, + "learning_rate": 0.0004197930207231119, + "loss": 3.774794101715088, + "step": 2775 + }, + { + "epoch": 38.5589519650655, + "grad_norm": 0.007018622010946274, + "learning_rate": 0.00041889328873129933, + "loss": 3.7112040519714355, + "step": 2776 + }, + { + "epoch": 38.57292576419214, + "grad_norm": 0.006895892322063446, + "learning_rate": 0.0004179944398528659, + "loss": 3.7497243881225586, + "step": 2777 + }, + { + "epoch": 38.58689956331878, + "grad_norm": 0.0071940235793590546, + "learning_rate": 0.0004170964748119979, + "loss": 3.7566123008728027, + "step": 2778 + }, + { + "epoch": 38.60087336244541, + "grad_norm": 0.006776201073080301, + "learning_rate": 0.0004161993943321702, + "loss": 3.7683968544006348, + "step": 2779 + }, + { + "epoch": 38.61484716157205, + "grad_norm": 0.006886148825287819, + "learning_rate": 0.0004153031991361428, + "loss": 3.759763240814209, + "step": 2780 + }, + { + "epoch": 38.62882096069869, + "grad_norm": 0.006466342136263847, + "learning_rate": 0.0004144078899459659, + "loss": 3.736006498336792, + "step": 2781 + }, + { + "epoch": 38.64279475982533, + "grad_norm": 0.007055724039673805, + "learning_rate": 0.0004135134674829725, + "loss": 3.70645809173584, + "step": 2782 + }, + { + "epoch": 38.65676855895197, + "grad_norm": 0.006948428228497505, + "learning_rate": 0.00041261993246778303, + "loss": 3.7365002632141113, + "step": 2783 + }, + { + "epoch": 38.670742358078606, + "grad_norm": 0.006491438951343298, + "learning_rate": 0.00041172728562030255, + "loss": 3.735001564025879, + "step": 2784 + }, + { + "epoch": 38.68471615720524, + "grad_norm": 0.006851755082607269, + "learning_rate": 0.0004108355276597209, + "loss": 3.720893383026123, + "step": 2785 + }, + { + "epoch": 38.698689956331876, + "grad_norm": 0.007266475819051266, + "learning_rate": 0.00040994465930451114, + "loss": 3.7746100425720215, + "step": 2786 + }, + { + "epoch": 38.712663755458514, + "grad_norm": 0.007112737279385328, + "learning_rate": 0.00040905468127242926, + "loss": 3.7441720962524414, + "step": 2787 + }, + { + "epoch": 38.72663755458515, + "grad_norm": 0.006514255423098803, + "learning_rate": 0.000408165594280515, + "loss": 3.7333898544311523, + "step": 2788 + }, + { + "epoch": 38.74061135371179, + "grad_norm": 0.007172788958996534, + "learning_rate": 0.0004072773990450896, + "loss": 3.7675576210021973, + "step": 2789 + }, + { + "epoch": 38.75458515283843, + "grad_norm": 0.00684684282168746, + "learning_rate": 0.0004063900962817564, + "loss": 3.7141308784484863, + "step": 2790 + }, + { + "epoch": 38.76855895196506, + "grad_norm": 0.006578145083039999, + "learning_rate": 0.00040550368670539777, + "loss": 3.7256317138671875, + "step": 2791 + }, + { + "epoch": 38.7825327510917, + "grad_norm": 0.007047492545098066, + "learning_rate": 0.00040461817103017926, + "loss": 3.7618556022644043, + "step": 2792 + }, + { + "epoch": 38.79650655021834, + "grad_norm": 0.007036720868200064, + "learning_rate": 0.00040373354996954487, + "loss": 3.7397944927215576, + "step": 2793 + }, + { + "epoch": 38.81048034934498, + "grad_norm": 0.00745626725256443, + "learning_rate": 0.00040284982423621714, + "loss": 3.7831153869628906, + "step": 2794 + }, + { + "epoch": 38.82445414847162, + "grad_norm": 0.007454817183315754, + "learning_rate": 0.00040196699454219797, + "loss": 3.7409911155700684, + "step": 2795 + }, + { + "epoch": 38.838427947598255, + "grad_norm": 0.0075650583021342754, + "learning_rate": 0.00040108506159876756, + "loss": 3.7074005603790283, + "step": 2796 + }, + { + "epoch": 38.852401746724894, + "grad_norm": 0.006908242590725422, + "learning_rate": 0.0004002040261164838, + "loss": 3.767744779586792, + "step": 2797 + }, + { + "epoch": 38.866375545851525, + "grad_norm": 0.0072857895866036415, + "learning_rate": 0.0003993238888051801, + "loss": 3.7349488735198975, + "step": 2798 + }, + { + "epoch": 38.880349344978164, + "grad_norm": 0.0072942771948874, + "learning_rate": 0.00039844465037396716, + "loss": 3.75968861579895, + "step": 2799 + }, + { + "epoch": 38.8943231441048, + "grad_norm": 0.0071496241725981236, + "learning_rate": 0.00039756631153123296, + "loss": 3.8237380981445312, + "step": 2800 + }, + { + "epoch": 38.90829694323144, + "grad_norm": 0.006803249940276146, + "learning_rate": 0.00039668887298463783, + "loss": 3.7468650341033936, + "step": 2801 + }, + { + "epoch": 38.92227074235808, + "grad_norm": 0.007433352991938591, + "learning_rate": 0.00039581233544111867, + "loss": 3.783236503601074, + "step": 2802 + }, + { + "epoch": 38.93624454148472, + "grad_norm": 0.006633558310568333, + "learning_rate": 0.0003949366996068858, + "loss": 3.75020694732666, + "step": 2803 + }, + { + "epoch": 38.95021834061135, + "grad_norm": 0.007415154483169317, + "learning_rate": 0.00039406196618742376, + "loss": 3.7545979022979736, + "step": 2804 + }, + { + "epoch": 38.96419213973799, + "grad_norm": 0.006997843272984028, + "learning_rate": 0.0003931881358874886, + "loss": 3.7404396533966064, + "step": 2805 + }, + { + "epoch": 38.97816593886463, + "grad_norm": 0.007417330518364906, + "learning_rate": 0.0003923152094111095, + "loss": 3.7575364112854004, + "step": 2806 + }, + { + "epoch": 38.992139737991266, + "grad_norm": 0.006936701480299234, + "learning_rate": 0.0003914431874615889, + "loss": 3.771845817565918, + "step": 2807 + }, + { + "epoch": 39.0, + "grad_norm": 0.008293208666145802, + "learning_rate": 0.00039057207074149795, + "loss": 3.746669292449951, + "step": 2808 + }, + { + "epoch": 39.0, + "eval_loss": 4.370476245880127, + "eval_runtime": 50.9368, + "eval_samples_per_second": 47.942, + "eval_steps_per_second": 1.512, + "step": 2808 + }, + { + "epoch": 39.01397379912664, + "grad_norm": 0.0077544888481497765, + "learning_rate": 0.0003897018599526799, + "loss": 3.6915597915649414, + "step": 2809 + }, + { + "epoch": 39.02794759825328, + "grad_norm": 0.008674043230712414, + "learning_rate": 0.000388832555796248, + "loss": 3.67698073387146, + "step": 2810 + }, + { + "epoch": 39.041921397379916, + "grad_norm": 0.00724575063213706, + "learning_rate": 0.00038796415897258523, + "loss": 3.6838502883911133, + "step": 2811 + }, + { + "epoch": 39.05589519650655, + "grad_norm": 0.00849415734410286, + "learning_rate": 0.00038709667018134246, + "loss": 3.69064998626709, + "step": 2812 + }, + { + "epoch": 39.069868995633186, + "grad_norm": 0.00769044877961278, + "learning_rate": 0.0003862300901214403, + "loss": 3.670502185821533, + "step": 2813 + }, + { + "epoch": 39.083842794759825, + "grad_norm": 0.007846124470233917, + "learning_rate": 0.00038536441949106647, + "loss": 3.767771005630493, + "step": 2814 + }, + { + "epoch": 39.09781659388646, + "grad_norm": 0.007555077783763409, + "learning_rate": 0.000384499658987676, + "loss": 3.7001633644104004, + "step": 2815 + }, + { + "epoch": 39.1117903930131, + "grad_norm": 0.007834251970052719, + "learning_rate": 0.0003836358093079915, + "loss": 3.704331874847412, + "step": 2816 + }, + { + "epoch": 39.12576419213974, + "grad_norm": 0.007608097977936268, + "learning_rate": 0.00038277287114799974, + "loss": 3.7108519077301025, + "step": 2817 + }, + { + "epoch": 39.13973799126637, + "grad_norm": 0.007420446258038282, + "learning_rate": 0.0003819108452029561, + "loss": 3.697843313217163, + "step": 2818 + }, + { + "epoch": 39.15371179039301, + "grad_norm": 0.008091001771390438, + "learning_rate": 0.000381049732167378, + "loss": 3.7380661964416504, + "step": 2819 + }, + { + "epoch": 39.16768558951965, + "grad_norm": 0.007265515625476837, + "learning_rate": 0.00038018953273504907, + "loss": 3.680748462677002, + "step": 2820 + }, + { + "epoch": 39.18165938864629, + "grad_norm": 0.007735357619822025, + "learning_rate": 0.00037933024759901636, + "loss": 3.6648077964782715, + "step": 2821 + }, + { + "epoch": 39.19563318777293, + "grad_norm": 0.008188895881175995, + "learning_rate": 0.0003784718774515907, + "loss": 3.706609010696411, + "step": 2822 + }, + { + "epoch": 39.209606986899566, + "grad_norm": 0.00816027820110321, + "learning_rate": 0.00037761442298434554, + "loss": 3.7368297576904297, + "step": 2823 + }, + { + "epoch": 39.223580786026204, + "grad_norm": 0.006931936834007502, + "learning_rate": 0.00037675788488811566, + "loss": 3.7121026515960693, + "step": 2824 + }, + { + "epoch": 39.237554585152836, + "grad_norm": 0.007615855894982815, + "learning_rate": 0.00037590226385299897, + "loss": 3.7379372119903564, + "step": 2825 + }, + { + "epoch": 39.251528384279474, + "grad_norm": 0.0073505607433617115, + "learning_rate": 0.0003750475605683535, + "loss": 3.7330894470214844, + "step": 2826 + }, + { + "epoch": 39.26550218340611, + "grad_norm": 0.007151979021728039, + "learning_rate": 0.00037419377572279897, + "loss": 3.713761568069458, + "step": 2827 + }, + { + "epoch": 39.27947598253275, + "grad_norm": 0.007310053799301386, + "learning_rate": 0.00037334091000421326, + "loss": 3.705451011657715, + "step": 2828 + }, + { + "epoch": 39.29344978165939, + "grad_norm": 0.0072968751192092896, + "learning_rate": 0.0003724889640997357, + "loss": 3.7550907135009766, + "step": 2829 + }, + { + "epoch": 39.30742358078603, + "grad_norm": 0.007197601720690727, + "learning_rate": 0.00037163793869576404, + "loss": 3.7346463203430176, + "step": 2830 + }, + { + "epoch": 39.32139737991266, + "grad_norm": 0.007823911495506763, + "learning_rate": 0.00037078783447795307, + "loss": 3.7257699966430664, + "step": 2831 + }, + { + "epoch": 39.3353711790393, + "grad_norm": 0.007118315435945988, + "learning_rate": 0.0003699386521312171, + "loss": 3.7014496326446533, + "step": 2832 + }, + { + "epoch": 39.34934497816594, + "grad_norm": 0.0069230347871780396, + "learning_rate": 0.00036909039233972666, + "loss": 3.6709237098693848, + "step": 2833 + }, + { + "epoch": 39.36331877729258, + "grad_norm": 0.007836844772100449, + "learning_rate": 0.00036824305578691, + "loss": 3.695725440979004, + "step": 2834 + }, + { + "epoch": 39.377292576419215, + "grad_norm": 0.007123398594558239, + "learning_rate": 0.00036739664315544956, + "loss": 3.7278594970703125, + "step": 2835 + }, + { + "epoch": 39.391266375545854, + "grad_norm": 0.007792143151164055, + "learning_rate": 0.00036655115512728586, + "loss": 3.7291927337646484, + "step": 2836 + }, + { + "epoch": 39.40524017467249, + "grad_norm": 0.007245976477861404, + "learning_rate": 0.00036570659238361345, + "loss": 3.7183046340942383, + "step": 2837 + }, + { + "epoch": 39.419213973799124, + "grad_norm": 0.007778054103255272, + "learning_rate": 0.0003648629556048804, + "loss": 3.7485761642456055, + "step": 2838 + }, + { + "epoch": 39.43318777292576, + "grad_norm": 0.007539471611380577, + "learning_rate": 0.00036402024547079007, + "loss": 3.7273335456848145, + "step": 2839 + }, + { + "epoch": 39.4471615720524, + "grad_norm": 0.007554007228463888, + "learning_rate": 0.0003631784626602987, + "loss": 3.684359073638916, + "step": 2840 + }, + { + "epoch": 39.46113537117904, + "grad_norm": 0.007862667553126812, + "learning_rate": 0.00036233760785161606, + "loss": 3.7223072052001953, + "step": 2841 + }, + { + "epoch": 39.47510917030568, + "grad_norm": 0.00869659148156643, + "learning_rate": 0.0003614976817222028, + "loss": 3.7475168704986572, + "step": 2842 + }, + { + "epoch": 39.48908296943232, + "grad_norm": 0.00696196174249053, + "learning_rate": 0.00036065868494877226, + "loss": 3.7773914337158203, + "step": 2843 + }, + { + "epoch": 39.50305676855895, + "grad_norm": 0.008952745236456394, + "learning_rate": 0.00035982061820729005, + "loss": 3.6761155128479004, + "step": 2844 + }, + { + "epoch": 39.51703056768559, + "grad_norm": 0.008341601118445396, + "learning_rate": 0.00035898348217297034, + "loss": 3.746553421020508, + "step": 2845 + }, + { + "epoch": 39.531004366812226, + "grad_norm": 0.007268146146088839, + "learning_rate": 0.00035814727752027885, + "loss": 3.6921701431274414, + "step": 2846 + }, + { + "epoch": 39.544978165938865, + "grad_norm": 0.007401647046208382, + "learning_rate": 0.00035731200492293064, + "loss": 3.722433090209961, + "step": 2847 + }, + { + "epoch": 39.5589519650655, + "grad_norm": 0.006969883106648922, + "learning_rate": 0.00035647766505389006, + "loss": 3.7449448108673096, + "step": 2848 + }, + { + "epoch": 39.57292576419214, + "grad_norm": 0.007616426330059767, + "learning_rate": 0.00035564425858536895, + "loss": 3.725113868713379, + "step": 2849 + }, + { + "epoch": 39.58689956331878, + "grad_norm": 0.006883042864501476, + "learning_rate": 0.00035481178618882814, + "loss": 3.717106342315674, + "step": 2850 + }, + { + "epoch": 39.60087336244541, + "grad_norm": 0.007106961216777563, + "learning_rate": 0.00035398024853497574, + "loss": 3.74045467376709, + "step": 2851 + }, + { + "epoch": 39.61484716157205, + "grad_norm": 0.007402827963232994, + "learning_rate": 0.00035314964629376664, + "loss": 3.7288095951080322, + "step": 2852 + }, + { + "epoch": 39.62882096069869, + "grad_norm": 0.006770852487534285, + "learning_rate": 0.00035231998013440227, + "loss": 3.7663369178771973, + "step": 2853 + }, + { + "epoch": 39.64279475982533, + "grad_norm": 0.007622968405485153, + "learning_rate": 0.0003514912507253287, + "loss": 3.747084140777588, + "step": 2854 + }, + { + "epoch": 39.65676855895197, + "grad_norm": 0.006876434665173292, + "learning_rate": 0.00035066345873423975, + "loss": 3.7591025829315186, + "step": 2855 + }, + { + "epoch": 39.670742358078606, + "grad_norm": 0.007071215659379959, + "learning_rate": 0.00034983660482807116, + "loss": 3.7583885192871094, + "step": 2856 + }, + { + "epoch": 39.68471615720524, + "grad_norm": 0.007130554411560297, + "learning_rate": 0.00034901068967300465, + "loss": 3.6950902938842773, + "step": 2857 + }, + { + "epoch": 39.698689956331876, + "grad_norm": 0.006981188431382179, + "learning_rate": 0.00034818571393446524, + "loss": 3.7390527725219727, + "step": 2858 + }, + { + "epoch": 39.712663755458514, + "grad_norm": 0.0067793820053339005, + "learning_rate": 0.0003473616782771209, + "loss": 3.703673839569092, + "step": 2859 + }, + { + "epoch": 39.72663755458515, + "grad_norm": 0.00674026133492589, + "learning_rate": 0.0003465385833648827, + "loss": 3.6988182067871094, + "step": 2860 + }, + { + "epoch": 39.74061135371179, + "grad_norm": 0.006553220096975565, + "learning_rate": 0.0003457164298609021, + "loss": 3.7367048263549805, + "step": 2861 + }, + { + "epoch": 39.75458515283843, + "grad_norm": 0.007079749833792448, + "learning_rate": 0.000344895218427575, + "loss": 3.773226022720337, + "step": 2862 + }, + { + "epoch": 39.76855895196506, + "grad_norm": 0.006091540213674307, + "learning_rate": 0.00034407494972653547, + "loss": 3.747929096221924, + "step": 2863 + }, + { + "epoch": 39.7825327510917, + "grad_norm": 0.006783034652471542, + "learning_rate": 0.00034325562441865935, + "loss": 3.7154481410980225, + "step": 2864 + }, + { + "epoch": 39.79650655021834, + "grad_norm": 0.006683558225631714, + "learning_rate": 0.0003424372431640625, + "loss": 3.751605987548828, + "step": 2865 + }, + { + "epoch": 39.81048034934498, + "grad_norm": 0.0066695502027869225, + "learning_rate": 0.0003416198066220997, + "loss": 3.7561278343200684, + "step": 2866 + }, + { + "epoch": 39.82445414847162, + "grad_norm": 0.006601935718208551, + "learning_rate": 0.0003408033154513655, + "loss": 3.754878520965576, + "step": 2867 + }, + { + "epoch": 39.838427947598255, + "grad_norm": 0.006911450996994972, + "learning_rate": 0.00033998777030969085, + "loss": 3.69533109664917, + "step": 2868 + }, + { + "epoch": 39.852401746724894, + "grad_norm": 0.0070345234125852585, + "learning_rate": 0.0003391731718541465, + "loss": 3.7328338623046875, + "step": 2869 + }, + { + "epoch": 39.866375545851525, + "grad_norm": 0.007238153368234634, + "learning_rate": 0.00033835952074103964, + "loss": 3.7370123863220215, + "step": 2870 + }, + { + "epoch": 39.880349344978164, + "grad_norm": 0.007295276038348675, + "learning_rate": 0.0003375468176259142, + "loss": 3.7123947143554688, + "step": 2871 + }, + { + "epoch": 39.8943231441048, + "grad_norm": 0.006471444386988878, + "learning_rate": 0.0003367350631635506, + "loss": 3.73347806930542, + "step": 2872 + }, + { + "epoch": 39.90829694323144, + "grad_norm": 0.007136075291782618, + "learning_rate": 0.00033592425800796447, + "loss": 3.7337799072265625, + "step": 2873 + }, + { + "epoch": 39.92227074235808, + "grad_norm": 0.0065771411173045635, + "learning_rate": 0.00033511440281240725, + "loss": 3.733506441116333, + "step": 2874 + }, + { + "epoch": 39.93624454148472, + "grad_norm": 0.006566311698406935, + "learning_rate": 0.0003343054982293643, + "loss": 3.7325632572174072, + "step": 2875 + }, + { + "epoch": 39.95021834061135, + "grad_norm": 0.006389300804585218, + "learning_rate": 0.00033349754491055535, + "loss": 3.7027206420898438, + "step": 2876 + }, + { + "epoch": 39.96419213973799, + "grad_norm": 0.006425539031624794, + "learning_rate": 0.00033269054350693386, + "loss": 3.751466989517212, + "step": 2877 + }, + { + "epoch": 39.97816593886463, + "grad_norm": 0.006735894829034805, + "learning_rate": 0.0003318844946686865, + "loss": 3.731363296508789, + "step": 2878 + }, + { + "epoch": 39.992139737991266, + "grad_norm": 0.0061791036278009415, + "learning_rate": 0.00033107939904523264, + "loss": 3.7592201232910156, + "step": 2879 + }, + { + "epoch": 40.0, + "grad_norm": 0.007923831231892109, + "learning_rate": 0.0003302752572852219, + "loss": 3.7247159481048584, + "step": 2880 + }, + { + "epoch": 40.0, + "eval_loss": 4.371238708496094, + "eval_runtime": 51.312, + "eval_samples_per_second": 47.591, + "eval_steps_per_second": 1.501, + "step": 2880 + }, + { + "epoch": 40.01397379912664, + "grad_norm": 0.007161229848861694, + "learning_rate": 0.0003294720700365385, + "loss": 3.6985998153686523, + "step": 2881 + }, + { + "epoch": 40.02794759825328, + "grad_norm": 0.007182417903095484, + "learning_rate": 0.00032866983794629457, + "loss": 3.7116494178771973, + "step": 2882 + }, + { + "epoch": 40.041921397379916, + "grad_norm": 0.0072847092524170876, + "learning_rate": 0.00032786856166083463, + "loss": 3.6759095191955566, + "step": 2883 + }, + { + "epoch": 40.05589519650655, + "grad_norm": 0.006694195792078972, + "learning_rate": 0.0003270682418257323, + "loss": 3.6635122299194336, + "step": 2884 + }, + { + "epoch": 40.069868995633186, + "grad_norm": 0.0073003629222512245, + "learning_rate": 0.0003262688790857914, + "loss": 3.7130517959594727, + "step": 2885 + }, + { + "epoch": 40.083842794759825, + "grad_norm": 0.006921815685927868, + "learning_rate": 0.0003254704740850432, + "loss": 3.6923139095306396, + "step": 2886 + }, + { + "epoch": 40.09781659388646, + "grad_norm": 0.007585546933114529, + "learning_rate": 0.00032467302746674805, + "loss": 3.680659770965576, + "step": 2887 + }, + { + "epoch": 40.1117903930131, + "grad_norm": 0.0069264741614460945, + "learning_rate": 0.0003238765398733956, + "loss": 3.7042646408081055, + "step": 2888 + }, + { + "epoch": 40.12576419213974, + "grad_norm": 0.007581264711916447, + "learning_rate": 0.0003230810119467003, + "loss": 3.6890318393707275, + "step": 2889 + }, + { + "epoch": 40.13973799126637, + "grad_norm": 0.007325596176087856, + "learning_rate": 0.00032228644432760544, + "loss": 3.728785753250122, + "step": 2890 + }, + { + "epoch": 40.15371179039301, + "grad_norm": 0.006920192390680313, + "learning_rate": 0.0003214928376562787, + "loss": 3.765744686126709, + "step": 2891 + }, + { + "epoch": 40.16768558951965, + "grad_norm": 0.007247452158480883, + "learning_rate": 0.00032070019257211596, + "loss": 3.699038505554199, + "step": 2892 + }, + { + "epoch": 40.18165938864629, + "grad_norm": 0.0073222690261900425, + "learning_rate": 0.0003199085097137361, + "loss": 3.6657447814941406, + "step": 2893 + }, + { + "epoch": 40.19563318777293, + "grad_norm": 0.0074598658829927444, + "learning_rate": 0.000319117789718984, + "loss": 3.6929068565368652, + "step": 2894 + }, + { + "epoch": 40.209606986899566, + "grad_norm": 0.007144949398934841, + "learning_rate": 0.0003183280332249286, + "loss": 3.7166645526885986, + "step": 2895 + }, + { + "epoch": 40.223580786026204, + "grad_norm": 0.0068069938570261, + "learning_rate": 0.0003175392408678621, + "loss": 3.650606155395508, + "step": 2896 + }, + { + "epoch": 40.237554585152836, + "grad_norm": 0.007320682052522898, + "learning_rate": 0.0003167514132833012, + "loss": 3.69647216796875, + "step": 2897 + }, + { + "epoch": 40.251528384279474, + "grad_norm": 0.007024327293038368, + "learning_rate": 0.00031596455110598264, + "loss": 3.6764140129089355, + "step": 2898 + }, + { + "epoch": 40.26550218340611, + "grad_norm": 0.007796750869601965, + "learning_rate": 0.0003151786549698689, + "loss": 3.6841444969177246, + "step": 2899 + }, + { + "epoch": 40.27947598253275, + "grad_norm": 0.007526555098593235, + "learning_rate": 0.0003143937255081411, + "loss": 3.7181622982025146, + "step": 2900 + }, + { + "epoch": 40.29344978165939, + "grad_norm": 0.00689823878929019, + "learning_rate": 0.0003136097633532032, + "loss": 3.6805999279022217, + "step": 2901 + }, + { + "epoch": 40.30742358078603, + "grad_norm": 0.007449387572705746, + "learning_rate": 0.00031282676913667903, + "loss": 3.719672679901123, + "step": 2902 + }, + { + "epoch": 40.32139737991266, + "grad_norm": 0.0072627803310751915, + "learning_rate": 0.00031204474348941336, + "loss": 3.692244291305542, + "step": 2903 + }, + { + "epoch": 40.3353711790393, + "grad_norm": 0.0073951175436377525, + "learning_rate": 0.00031126368704147026, + "loss": 3.734830379486084, + "step": 2904 + }, + { + "epoch": 40.34934497816594, + "grad_norm": 0.007486097048968077, + "learning_rate": 0.0003104836004221318, + "loss": 3.6959404945373535, + "step": 2905 + }, + { + "epoch": 40.36331877729258, + "grad_norm": 0.007235408294945955, + "learning_rate": 0.0003097044842599005, + "loss": 3.714496612548828, + "step": 2906 + }, + { + "epoch": 40.377292576419215, + "grad_norm": 0.0074761975556612015, + "learning_rate": 0.00030892633918249584, + "loss": 3.6801066398620605, + "step": 2907 + }, + { + "epoch": 40.391266375545854, + "grad_norm": 0.007354153320193291, + "learning_rate": 0.0003081491658168555, + "loss": 3.7053956985473633, + "step": 2908 + }, + { + "epoch": 40.40524017467249, + "grad_norm": 0.0077071161940693855, + "learning_rate": 0.0003073729647891338, + "loss": 3.724217653274536, + "step": 2909 + }, + { + "epoch": 40.419213973799124, + "grad_norm": 0.007286814507097006, + "learning_rate": 0.00030659773672470227, + "loss": 3.717681646347046, + "step": 2910 + }, + { + "epoch": 40.43318777292576, + "grad_norm": 0.00741427019238472, + "learning_rate": 0.00030582348224814803, + "loss": 3.7178781032562256, + "step": 2911 + }, + { + "epoch": 40.4471615720524, + "grad_norm": 0.007106068544089794, + "learning_rate": 0.00030505020198327355, + "loss": 3.671164035797119, + "step": 2912 + }, + { + "epoch": 40.46113537117904, + "grad_norm": 0.006869092117995024, + "learning_rate": 0.0003042778965530969, + "loss": 3.7310664653778076, + "step": 2913 + }, + { + "epoch": 40.47510917030568, + "grad_norm": 0.00765156140550971, + "learning_rate": 0.0003035065665798508, + "loss": 3.7417078018188477, + "step": 2914 + }, + { + "epoch": 40.48908296943232, + "grad_norm": 0.006932055577635765, + "learning_rate": 0.00030273621268498183, + "loss": 3.6812591552734375, + "step": 2915 + }, + { + "epoch": 40.50305676855895, + "grad_norm": 0.008092030882835388, + "learning_rate": 0.00030196683548915055, + "loss": 3.69986891746521, + "step": 2916 + }, + { + "epoch": 40.51703056768559, + "grad_norm": 0.00719967158511281, + "learning_rate": 0.00030119843561222915, + "loss": 3.7247209548950195, + "step": 2917 + }, + { + "epoch": 40.531004366812226, + "grad_norm": 0.007408039178699255, + "learning_rate": 0.00030043101367330504, + "loss": 3.7106001377105713, + "step": 2918 + }, + { + "epoch": 40.544978165938865, + "grad_norm": 0.007386322598904371, + "learning_rate": 0.00029966457029067503, + "loss": 3.7073440551757812, + "step": 2919 + }, + { + "epoch": 40.5589519650655, + "grad_norm": 0.007485459093004465, + "learning_rate": 0.00029889910608184924, + "loss": 3.730858325958252, + "step": 2920 + }, + { + "epoch": 40.57292576419214, + "grad_norm": 0.007635290268808603, + "learning_rate": 0.00029813462166354814, + "loss": 3.6909053325653076, + "step": 2921 + }, + { + "epoch": 40.58689956331878, + "grad_norm": 0.0069068302400410175, + "learning_rate": 0.0002973711176517032, + "loss": 3.7286834716796875, + "step": 2922 + }, + { + "epoch": 40.60087336244541, + "grad_norm": 0.00698851840570569, + "learning_rate": 0.0002966085946614562, + "loss": 3.6587166786193848, + "step": 2923 + }, + { + "epoch": 40.61484716157205, + "grad_norm": 0.007312684785574675, + "learning_rate": 0.0002958470533071569, + "loss": 3.716031789779663, + "step": 2924 + }, + { + "epoch": 40.62882096069869, + "grad_norm": 0.006896916776895523, + "learning_rate": 0.00029508649420236687, + "loss": 3.7105724811553955, + "step": 2925 + }, + { + "epoch": 40.64279475982533, + "grad_norm": 0.007188099902123213, + "learning_rate": 0.00029432691795985394, + "loss": 3.712202548980713, + "step": 2926 + }, + { + "epoch": 40.65676855895197, + "grad_norm": 0.006414467468857765, + "learning_rate": 0.000293568325191595, + "loss": 3.71462345123291, + "step": 2927 + }, + { + "epoch": 40.670742358078606, + "grad_norm": 0.006734516471624374, + "learning_rate": 0.00029281071650877496, + "loss": 3.7277159690856934, + "step": 2928 + }, + { + "epoch": 40.68471615720524, + "grad_norm": 0.0067634680308401585, + "learning_rate": 0.00029205409252178524, + "loss": 3.6998372077941895, + "step": 2929 + }, + { + "epoch": 40.698689956331876, + "grad_norm": 0.006864939816296101, + "learning_rate": 0.00029129845384022446, + "loss": 3.7656338214874268, + "step": 2930 + }, + { + "epoch": 40.712663755458514, + "grad_norm": 0.006820365786552429, + "learning_rate": 0.00029054380107289626, + "loss": 3.7018871307373047, + "step": 2931 + }, + { + "epoch": 40.72663755458515, + "grad_norm": 0.006954800803214312, + "learning_rate": 0.00028979013482781114, + "loss": 3.7410342693328857, + "step": 2932 + }, + { + "epoch": 40.74061135371179, + "grad_norm": 0.0072957612574100494, + "learning_rate": 0.00028903745571218453, + "loss": 3.6964735984802246, + "step": 2933 + }, + { + "epoch": 40.75458515283843, + "grad_norm": 0.006805056240409613, + "learning_rate": 0.00028828576433243675, + "loss": 3.6988587379455566, + "step": 2934 + }, + { + "epoch": 40.76855895196506, + "grad_norm": 0.007712861057370901, + "learning_rate": 0.0002875350612941908, + "loss": 3.73240327835083, + "step": 2935 + }, + { + "epoch": 40.7825327510917, + "grad_norm": 0.006683751940727234, + "learning_rate": 0.0002867853472022755, + "loss": 3.6936707496643066, + "step": 2936 + }, + { + "epoch": 40.79650655021834, + "grad_norm": 0.007726171053946018, + "learning_rate": 0.0002860366226607223, + "loss": 3.7266287803649902, + "step": 2937 + }, + { + "epoch": 40.81048034934498, + "grad_norm": 0.007187390234321356, + "learning_rate": 0.00028528888827276396, + "loss": 3.7381067276000977, + "step": 2938 + }, + { + "epoch": 40.82445414847162, + "grad_norm": 0.007067828439176083, + "learning_rate": 0.00028454214464083706, + "loss": 3.71884822845459, + "step": 2939 + }, + { + "epoch": 40.838427947598255, + "grad_norm": 0.007044029422104359, + "learning_rate": 0.00028379639236657966, + "loss": 3.6823341846466064, + "step": 2940 + }, + { + "epoch": 40.852401746724894, + "grad_norm": 0.007027678657323122, + "learning_rate": 0.000283051632050831, + "loss": 3.7312417030334473, + "step": 2941 + }, + { + "epoch": 40.866375545851525, + "grad_norm": 0.0076786368153989315, + "learning_rate": 0.00028230786429363046, + "loss": 3.775012493133545, + "step": 2942 + }, + { + "epoch": 40.880349344978164, + "grad_norm": 0.0067380438558757305, + "learning_rate": 0.000281565089694218, + "loss": 3.7203941345214844, + "step": 2943 + }, + { + "epoch": 40.8943231441048, + "grad_norm": 0.007500625215470791, + "learning_rate": 0.00028082330885103516, + "loss": 3.7320876121520996, + "step": 2944 + }, + { + "epoch": 40.90829694323144, + "grad_norm": 0.006601814646273851, + "learning_rate": 0.00028008252236172003, + "loss": 3.7289505004882812, + "step": 2945 + }, + { + "epoch": 40.92227074235808, + "grad_norm": 0.00755319744348526, + "learning_rate": 0.00027934273082311147, + "loss": 3.7076311111450195, + "step": 2946 + }, + { + "epoch": 40.93624454148472, + "grad_norm": 0.0068661835975945, + "learning_rate": 0.0002786039348312459, + "loss": 3.6950671672821045, + "step": 2947 + }, + { + "epoch": 40.95021834061135, + "grad_norm": 0.007409265730530024, + "learning_rate": 0.00027786613498135843, + "loss": 3.7798776626586914, + "step": 2948 + }, + { + "epoch": 40.96419213973799, + "grad_norm": 0.006714935880154371, + "learning_rate": 0.0002771293318678804, + "loss": 3.690377712249756, + "step": 2949 + }, + { + "epoch": 40.97816593886463, + "grad_norm": 0.007498119492083788, + "learning_rate": 0.00027639352608444087, + "loss": 3.741222381591797, + "step": 2950 + }, + { + "epoch": 40.992139737991266, + "grad_norm": 0.006700210273265839, + "learning_rate": 0.0002756587182238656, + "loss": 3.7065296173095703, + "step": 2951 + }, + { + "epoch": 41.0, + "grad_norm": 0.008518884889781475, + "learning_rate": 0.00027492490887817597, + "loss": 3.7010321617126465, + "step": 2952 + }, + { + "epoch": 41.0, + "eval_loss": 4.372326374053955, + "eval_runtime": 48.97, + "eval_samples_per_second": 49.867, + "eval_steps_per_second": 1.572, + "step": 2952 + }, + { + "epoch": 41.01397379912664, + "grad_norm": 0.007463657297194004, + "learning_rate": 0.00027419209863858865, + "loss": 3.6893815994262695, + "step": 2953 + }, + { + "epoch": 41.02794759825328, + "grad_norm": 0.0069931019097566605, + "learning_rate": 0.00027346028809551576, + "loss": 3.6620845794677734, + "step": 2954 + }, + { + "epoch": 41.041921397379916, + "grad_norm": 0.0076762037351727486, + "learning_rate": 0.00027272947783856433, + "loss": 3.715946674346924, + "step": 2955 + }, + { + "epoch": 41.05589519650655, + "grad_norm": 0.007143040653318167, + "learning_rate": 0.0002719996684565341, + "loss": 3.67921781539917, + "step": 2956 + }, + { + "epoch": 41.069868995633186, + "grad_norm": 0.007577868644148111, + "learning_rate": 0.00027127086053741964, + "loss": 3.6886801719665527, + "step": 2957 + }, + { + "epoch": 41.083842794759825, + "grad_norm": 0.007018061354756355, + "learning_rate": 0.0002705430546684085, + "loss": 3.673877716064453, + "step": 2958 + }, + { + "epoch": 41.09781659388646, + "grad_norm": 0.0072300247848033905, + "learning_rate": 0.0002698162514358805, + "loss": 3.730193614959717, + "step": 2959 + }, + { + "epoch": 41.1117903930131, + "grad_norm": 0.00714140385389328, + "learning_rate": 0.00026909045142540843, + "loss": 3.6612749099731445, + "step": 2960 + }, + { + "epoch": 41.12576419213974, + "grad_norm": 0.0071529485285282135, + "learning_rate": 0.0002683656552217553, + "loss": 3.641969680786133, + "step": 2961 + }, + { + "epoch": 41.13973799126637, + "grad_norm": 0.00726350536569953, + "learning_rate": 0.0002676418634088772, + "loss": 3.697068452835083, + "step": 2962 + }, + { + "epoch": 41.15371179039301, + "grad_norm": 0.007082062773406506, + "learning_rate": 0.00026691907656991963, + "loss": 3.6732699871063232, + "step": 2963 + }, + { + "epoch": 41.16768558951965, + "grad_norm": 0.0069284942001104355, + "learning_rate": 0.00026619729528721897, + "loss": 3.6701955795288086, + "step": 2964 + }, + { + "epoch": 41.18165938864629, + "grad_norm": 0.007283426355570555, + "learning_rate": 0.00026547652014230147, + "loss": 3.6422624588012695, + "step": 2965 + }, + { + "epoch": 41.19563318777293, + "grad_norm": 0.006717352196574211, + "learning_rate": 0.0002647567517158825, + "loss": 3.6600942611694336, + "step": 2966 + }, + { + "epoch": 41.209606986899566, + "grad_norm": 0.006842337548732758, + "learning_rate": 0.00026403799058786695, + "loss": 3.690095901489258, + "step": 2967 + }, + { + "epoch": 41.223580786026204, + "grad_norm": 0.006624994333833456, + "learning_rate": 0.000263320237337347, + "loss": 3.7145729064941406, + "step": 2968 + }, + { + "epoch": 41.237554585152836, + "grad_norm": 0.006822563707828522, + "learning_rate": 0.0002626034925426036, + "loss": 3.734589099884033, + "step": 2969 + }, + { + "epoch": 41.251528384279474, + "grad_norm": 0.006343246903270483, + "learning_rate": 0.0002618877567811054, + "loss": 3.676018238067627, + "step": 2970 + }, + { + "epoch": 41.26550218340611, + "grad_norm": 0.007226140238344669, + "learning_rate": 0.00026117303062950795, + "loss": 3.705630302429199, + "step": 2971 + }, + { + "epoch": 41.27947598253275, + "grad_norm": 0.0066663846373558044, + "learning_rate": 0.0002604593146636522, + "loss": 3.707524299621582, + "step": 2972 + }, + { + "epoch": 41.29344978165939, + "grad_norm": 0.007017293944954872, + "learning_rate": 0.00025974660945856726, + "loss": 3.654975652694702, + "step": 2973 + }, + { + "epoch": 41.30742358078603, + "grad_norm": 0.006494272034615278, + "learning_rate": 0.0002590349155884669, + "loss": 3.7084226608276367, + "step": 2974 + }, + { + "epoch": 41.32139737991266, + "grad_norm": 0.007430107332766056, + "learning_rate": 0.0002583242336267492, + "loss": 3.681856393814087, + "step": 2975 + }, + { + "epoch": 41.3353711790393, + "grad_norm": 0.0065765841864049435, + "learning_rate": 0.00025761456414599807, + "loss": 3.7148854732513428, + "step": 2976 + }, + { + "epoch": 41.34934497816594, + "grad_norm": 0.00703916372731328, + "learning_rate": 0.00025690590771798165, + "loss": 3.707977294921875, + "step": 2977 + }, + { + "epoch": 41.36331877729258, + "grad_norm": 0.006881422363221645, + "learning_rate": 0.0002561982649136517, + "loss": 3.6452829837799072, + "step": 2978 + }, + { + "epoch": 41.377292576419215, + "grad_norm": 0.006829532328993082, + "learning_rate": 0.00025549163630314233, + "loss": 3.6812632083892822, + "step": 2979 + }, + { + "epoch": 41.391266375545854, + "grad_norm": 0.0069242967292666435, + "learning_rate": 0.00025478602245577255, + "loss": 3.6699113845825195, + "step": 2980 + }, + { + "epoch": 41.40524017467249, + "grad_norm": 0.00673352275043726, + "learning_rate": 0.00025408142394004265, + "loss": 3.67539644241333, + "step": 2981 + }, + { + "epoch": 41.419213973799124, + "grad_norm": 0.006906552705913782, + "learning_rate": 0.0002533778413236342, + "loss": 3.695683479309082, + "step": 2982 + }, + { + "epoch": 41.43318777292576, + "grad_norm": 0.007096703629940748, + "learning_rate": 0.0002526752751734114, + "loss": 3.720625877380371, + "step": 2983 + }, + { + "epoch": 41.4471615720524, + "grad_norm": 0.006843909155577421, + "learning_rate": 0.000251973726055419, + "loss": 3.6780426502227783, + "step": 2984 + }, + { + "epoch": 41.46113537117904, + "grad_norm": 0.006776071153581142, + "learning_rate": 0.00025127319453488277, + "loss": 3.702991485595703, + "step": 2985 + }, + { + "epoch": 41.47510917030568, + "grad_norm": 0.006684580352157354, + "learning_rate": 0.00025057368117620797, + "loss": 3.7161166667938232, + "step": 2986 + }, + { + "epoch": 41.48908296943232, + "grad_norm": 0.006576182786375284, + "learning_rate": 0.00024987518654297935, + "loss": 3.711947202682495, + "step": 2987 + }, + { + "epoch": 41.50305676855895, + "grad_norm": 0.006553746294230223, + "learning_rate": 0.0002491777111979629, + "loss": 3.699580669403076, + "step": 2988 + }, + { + "epoch": 41.51703056768559, + "grad_norm": 0.0063622179441154, + "learning_rate": 0.00024848125570310067, + "loss": 3.660463809967041, + "step": 2989 + }, + { + "epoch": 41.531004366812226, + "grad_norm": 0.006369025446474552, + "learning_rate": 0.00024778582061951466, + "loss": 3.712718963623047, + "step": 2990 + }, + { + "epoch": 41.544978165938865, + "grad_norm": 0.0066086845472455025, + "learning_rate": 0.00024709140650750426, + "loss": 3.7337706089019775, + "step": 2991 + }, + { + "epoch": 41.5589519650655, + "grad_norm": 0.006698573008179665, + "learning_rate": 0.00024639801392654687, + "loss": 3.693399429321289, + "step": 2992 + }, + { + "epoch": 41.57292576419214, + "grad_norm": 0.006572197191417217, + "learning_rate": 0.0002457056434352956, + "loss": 3.711644411087036, + "step": 2993 + }, + { + "epoch": 41.58689956331878, + "grad_norm": 0.0066736009903252125, + "learning_rate": 0.0002450142955915812, + "loss": 3.706545352935791, + "step": 2994 + }, + { + "epoch": 41.60087336244541, + "grad_norm": 0.006779580842703581, + "learning_rate": 0.0002443239709524101, + "loss": 3.6896963119506836, + "step": 2995 + }, + { + "epoch": 41.61484716157205, + "grad_norm": 0.0064375437796115875, + "learning_rate": 0.0002436346700739646, + "loss": 3.646042823791504, + "step": 2996 + }, + { + "epoch": 41.62882096069869, + "grad_norm": 0.006759198382496834, + "learning_rate": 0.00024294639351160198, + "loss": 3.6894006729125977, + "step": 2997 + }, + { + "epoch": 41.64279475982533, + "grad_norm": 0.00644648214802146, + "learning_rate": 0.00024225914181985343, + "loss": 3.7088422775268555, + "step": 2998 + }, + { + "epoch": 41.65676855895197, + "grad_norm": 0.00702979601919651, + "learning_rate": 0.00024157291555242663, + "loss": 3.705204486846924, + "step": 2999 + }, + { + "epoch": 41.670742358078606, + "grad_norm": 0.006393326912075281, + "learning_rate": 0.00024088771526220045, + "loss": 3.738778829574585, + "step": 3000 + }, + { + "epoch": 41.68471615720524, + "grad_norm": 0.006522264331579208, + "learning_rate": 0.00024020354150122884, + "loss": 3.7124788761138916, + "step": 3001 + }, + { + "epoch": 41.698689956331876, + "grad_norm": 0.006918660830706358, + "learning_rate": 0.0002395203948207383, + "loss": 3.703273296356201, + "step": 3002 + }, + { + "epoch": 41.712663755458514, + "grad_norm": 0.0067405398003757, + "learning_rate": 0.00023883827577112786, + "loss": 3.735112190246582, + "step": 3003 + }, + { + "epoch": 41.72663755458515, + "grad_norm": 0.006900794804096222, + "learning_rate": 0.00023815718490196856, + "loss": 3.707573413848877, + "step": 3004 + }, + { + "epoch": 41.74061135371179, + "grad_norm": 0.006647802423685789, + "learning_rate": 0.0002374771227620022, + "loss": 3.7194061279296875, + "step": 3005 + }, + { + "epoch": 41.75458515283843, + "grad_norm": 0.006786358542740345, + "learning_rate": 0.00023679808989914383, + "loss": 3.684828758239746, + "step": 3006 + }, + { + "epoch": 41.76855895196506, + "grad_norm": 0.006440348923206329, + "learning_rate": 0.0002361200868604771, + "loss": 3.679628610610962, + "step": 3007 + }, + { + "epoch": 41.7825327510917, + "grad_norm": 0.006847103592008352, + "learning_rate": 0.0002354431141922572, + "loss": 3.7070424556732178, + "step": 3008 + }, + { + "epoch": 41.79650655021834, + "grad_norm": 0.00643899803981185, + "learning_rate": 0.00023476717243990815, + "loss": 3.7474443912506104, + "step": 3009 + }, + { + "epoch": 41.81048034934498, + "grad_norm": 0.0070899054408073425, + "learning_rate": 0.00023409226214802462, + "loss": 3.6964569091796875, + "step": 3010 + }, + { + "epoch": 41.82445414847162, + "grad_norm": 0.006639422383159399, + "learning_rate": 0.00023341838386036994, + "loss": 3.7240355014801025, + "step": 3011 + }, + { + "epoch": 41.838427947598255, + "grad_norm": 0.0067732552997767925, + "learning_rate": 0.000232745538119875, + "loss": 3.691311836242676, + "step": 3012 + }, + { + "epoch": 41.852401746724894, + "grad_norm": 0.006991941947489977, + "learning_rate": 0.0002320737254686396, + "loss": 3.703491449356079, + "step": 3013 + }, + { + "epoch": 41.866375545851525, + "grad_norm": 0.006394840776920319, + "learning_rate": 0.0002314029464479313, + "loss": 3.6963045597076416, + "step": 3014 + }, + { + "epoch": 41.880349344978164, + "grad_norm": 0.00672922981902957, + "learning_rate": 0.00023073320159818463, + "loss": 3.6913256645202637, + "step": 3015 + }, + { + "epoch": 41.8943231441048, + "grad_norm": 0.006367729511111975, + "learning_rate": 0.00023006449145900022, + "loss": 3.712766170501709, + "step": 3016 + }, + { + "epoch": 41.90829694323144, + "grad_norm": 0.006370930001139641, + "learning_rate": 0.0002293968165691467, + "loss": 3.683473587036133, + "step": 3017 + }, + { + "epoch": 41.92227074235808, + "grad_norm": 0.007074430584907532, + "learning_rate": 0.00022873017746655758, + "loss": 3.7364859580993652, + "step": 3018 + }, + { + "epoch": 41.93624454148472, + "grad_norm": 0.00689726835116744, + "learning_rate": 0.00022806457468833126, + "loss": 3.7360360622406006, + "step": 3019 + }, + { + "epoch": 41.95021834061135, + "grad_norm": 0.006516663823276758, + "learning_rate": 0.0002274000087707323, + "loss": 3.7183632850646973, + "step": 3020 + }, + { + "epoch": 41.96419213973799, + "grad_norm": 0.006655904930084944, + "learning_rate": 0.00022673648024918938, + "loss": 3.7091434001922607, + "step": 3021 + }, + { + "epoch": 41.97816593886463, + "grad_norm": 0.006546393036842346, + "learning_rate": 0.00022607398965829604, + "loss": 3.682645320892334, + "step": 3022 + }, + { + "epoch": 41.992139737991266, + "grad_norm": 0.006579853594303131, + "learning_rate": 0.0002254125375318079, + "loss": 3.707092761993408, + "step": 3023 + }, + { + "epoch": 42.0, + "grad_norm": 0.007749224081635475, + "learning_rate": 0.00022475212440264524, + "loss": 3.6835598945617676, + "step": 3024 + }, + { + "epoch": 42.0, + "eval_loss": 4.3733673095703125, + "eval_runtime": 54.197, + "eval_samples_per_second": 45.058, + "eval_steps_per_second": 1.421, + "step": 3024 + }, + { + "epoch": 42.01397379912664, + "grad_norm": 0.006405573338270187, + "learning_rate": 0.000224092750802892, + "loss": 3.7344212532043457, + "step": 3025 + }, + { + "epoch": 42.02794759825328, + "grad_norm": 0.006732994224876165, + "learning_rate": 0.0002234344172637924, + "loss": 3.684095621109009, + "step": 3026 + }, + { + "epoch": 42.041921397379916, + "grad_norm": 0.00679958239197731, + "learning_rate": 0.00022277712431575434, + "loss": 3.697983741760254, + "step": 3027 + }, + { + "epoch": 42.05589519650655, + "grad_norm": 0.00670752814039588, + "learning_rate": 0.00022212087248834695, + "loss": 3.6364173889160156, + "step": 3028 + }, + { + "epoch": 42.069868995633186, + "grad_norm": 0.006356659810990095, + "learning_rate": 0.00022146566231030071, + "loss": 3.642988681793213, + "step": 3029 + }, + { + "epoch": 42.083842794759825, + "grad_norm": 0.006624842528253794, + "learning_rate": 0.000220811494309506, + "loss": 3.6682677268981934, + "step": 3030 + }, + { + "epoch": 42.09781659388646, + "grad_norm": 0.006486820988357067, + "learning_rate": 0.00022015836901301408, + "loss": 3.6703057289123535, + "step": 3031 + }, + { + "epoch": 42.1117903930131, + "grad_norm": 0.006477998569607735, + "learning_rate": 0.0002195062869470374, + "loss": 3.6891942024230957, + "step": 3032 + }, + { + "epoch": 42.12576419213974, + "grad_norm": 0.006716342642903328, + "learning_rate": 0.0002188552486369454, + "loss": 3.6837644577026367, + "step": 3033 + }, + { + "epoch": 42.13973799126637, + "grad_norm": 0.006492446642369032, + "learning_rate": 0.00021820525460726888, + "loss": 3.66949462890625, + "step": 3034 + }, + { + "epoch": 42.15371179039301, + "grad_norm": 0.0064978403970599174, + "learning_rate": 0.00021755630538169502, + "loss": 3.656477451324463, + "step": 3035 + }, + { + "epoch": 42.16768558951965, + "grad_norm": 0.006488895509392023, + "learning_rate": 0.000216908401483072, + "loss": 3.683609962463379, + "step": 3036 + }, + { + "epoch": 42.18165938864629, + "grad_norm": 0.0063864924013614655, + "learning_rate": 0.00021626154343340316, + "loss": 3.6635069847106934, + "step": 3037 + }, + { + "epoch": 42.19563318777293, + "grad_norm": 0.006423276849091053, + "learning_rate": 0.00021561573175385077, + "loss": 3.675956964492798, + "step": 3038 + }, + { + "epoch": 42.209606986899566, + "grad_norm": 0.006375681143254042, + "learning_rate": 0.00021497096696473364, + "loss": 3.662991523742676, + "step": 3039 + }, + { + "epoch": 42.223580786026204, + "grad_norm": 0.006157599855214357, + "learning_rate": 0.00021432724958552707, + "loss": 3.685154438018799, + "step": 3040 + }, + { + "epoch": 42.237554585152836, + "grad_norm": 0.0064014773815870285, + "learning_rate": 0.0002136845801348629, + "loss": 3.6482927799224854, + "step": 3041 + }, + { + "epoch": 42.251528384279474, + "grad_norm": 0.006111831869930029, + "learning_rate": 0.0002130429591305271, + "loss": 3.6458590030670166, + "step": 3042 + }, + { + "epoch": 42.26550218340611, + "grad_norm": 0.0062011126428842545, + "learning_rate": 0.0002124023870894639, + "loss": 3.6651253700256348, + "step": 3043 + }, + { + "epoch": 42.27947598253275, + "grad_norm": 0.006603785324841738, + "learning_rate": 0.00021176286452776909, + "loss": 3.6931943893432617, + "step": 3044 + }, + { + "epoch": 42.29344978165939, + "grad_norm": 0.006363562773913145, + "learning_rate": 0.00021112439196069516, + "loss": 3.7063207626342773, + "step": 3045 + }, + { + "epoch": 42.30742358078603, + "grad_norm": 0.006258734501898289, + "learning_rate": 0.0002104869699026478, + "loss": 3.689711093902588, + "step": 3046 + }, + { + "epoch": 42.32139737991266, + "grad_norm": 0.006544198840856552, + "learning_rate": 0.00020985059886718659, + "loss": 3.6803956031799316, + "step": 3047 + }, + { + "epoch": 42.3353711790393, + "grad_norm": 0.006236413959413767, + "learning_rate": 0.00020921527936702418, + "loss": 3.681662082672119, + "step": 3048 + }, + { + "epoch": 42.34934497816594, + "grad_norm": 0.006620924919843674, + "learning_rate": 0.00020858101191402557, + "loss": 3.6708180904388428, + "step": 3049 + }, + { + "epoch": 42.36331877729258, + "grad_norm": 0.006312021054327488, + "learning_rate": 0.00020794779701920895, + "loss": 3.68668270111084, + "step": 3050 + }, + { + "epoch": 42.377292576419215, + "grad_norm": 0.006595146376639605, + "learning_rate": 0.00020731563519274407, + "loss": 3.709224224090576, + "step": 3051 + }, + { + "epoch": 42.391266375545854, + "grad_norm": 0.006226524710655212, + "learning_rate": 0.00020668452694395234, + "loss": 3.688267707824707, + "step": 3052 + }, + { + "epoch": 42.40524017467249, + "grad_norm": 0.006482335273176432, + "learning_rate": 0.00020605447278130616, + "loss": 3.6906940937042236, + "step": 3053 + }, + { + "epoch": 42.419213973799124, + "grad_norm": 0.006389188580214977, + "learning_rate": 0.00020542547321242892, + "loss": 3.6813313961029053, + "step": 3054 + }, + { + "epoch": 42.43318777292576, + "grad_norm": 0.006554862484335899, + "learning_rate": 0.00020479752874409457, + "loss": 3.645631790161133, + "step": 3055 + }, + { + "epoch": 42.4471615720524, + "grad_norm": 0.00646898802369833, + "learning_rate": 0.0002041706398822257, + "loss": 3.691580295562744, + "step": 3056 + }, + { + "epoch": 42.46113537117904, + "grad_norm": 0.006872115656733513, + "learning_rate": 0.0002035448071318961, + "loss": 3.716787815093994, + "step": 3057 + }, + { + "epoch": 42.47510917030568, + "grad_norm": 0.006251885090023279, + "learning_rate": 0.00020292003099732744, + "loss": 3.666078567504883, + "step": 3058 + }, + { + "epoch": 42.48908296943232, + "grad_norm": 0.006771737243980169, + "learning_rate": 0.00020229631198189084, + "loss": 3.621438503265381, + "step": 3059 + }, + { + "epoch": 42.50305676855895, + "grad_norm": 0.00606488948687911, + "learning_rate": 0.0002016736505881058, + "loss": 3.6710901260375977, + "step": 3060 + }, + { + "epoch": 42.51703056768559, + "grad_norm": 0.006203590892255306, + "learning_rate": 0.00020105204731763832, + "loss": 3.6667685508728027, + "step": 3061 + }, + { + "epoch": 42.531004366812226, + "grad_norm": 0.006625751499086618, + "learning_rate": 0.00020043150267130412, + "loss": 3.648806095123291, + "step": 3062 + }, + { + "epoch": 42.544978165938865, + "grad_norm": 0.006358640734106302, + "learning_rate": 0.000199812017149064, + "loss": 3.7056827545166016, + "step": 3063 + }, + { + "epoch": 42.5589519650655, + "grad_norm": 0.006539507769048214, + "learning_rate": 0.00019919359125002647, + "loss": 3.6906323432922363, + "step": 3064 + }, + { + "epoch": 42.57292576419214, + "grad_norm": 0.007050348911434412, + "learning_rate": 0.00019857622547244627, + "loss": 3.6898245811462402, + "step": 3065 + }, + { + "epoch": 42.58689956331878, + "grad_norm": 0.006209646817296743, + "learning_rate": 0.00019795992031372358, + "loss": 3.6794214248657227, + "step": 3066 + }, + { + "epoch": 42.60087336244541, + "grad_norm": 0.0064559150487184525, + "learning_rate": 0.00019734467627040487, + "loss": 3.6937336921691895, + "step": 3067 + }, + { + "epoch": 42.61484716157205, + "grad_norm": 0.006774613168090582, + "learning_rate": 0.00019673049383818025, + "loss": 3.7005720138549805, + "step": 3068 + }, + { + "epoch": 42.62882096069869, + "grad_norm": 0.0061691561713814735, + "learning_rate": 0.00019611737351188642, + "loss": 3.7170605659484863, + "step": 3069 + }, + { + "epoch": 42.64279475982533, + "grad_norm": 0.0070815603248775005, + "learning_rate": 0.0001955053157855027, + "loss": 3.722456693649292, + "step": 3070 + }, + { + "epoch": 42.65676855895197, + "grad_norm": 0.006411026697605848, + "learning_rate": 0.00019489432115215308, + "loss": 3.669548511505127, + "step": 3071 + }, + { + "epoch": 42.670742358078606, + "grad_norm": 0.006817718036472797, + "learning_rate": 0.00019428439010410492, + "loss": 3.705488681793213, + "step": 3072 + }, + { + "epoch": 42.68471615720524, + "grad_norm": 0.006276566535234451, + "learning_rate": 0.0001936755231327688, + "loss": 3.7110562324523926, + "step": 3073 + }, + { + "epoch": 42.698689956331876, + "grad_norm": 0.006713696289807558, + "learning_rate": 0.0001930677207286977, + "loss": 3.7030649185180664, + "step": 3074 + }, + { + "epoch": 42.712663755458514, + "grad_norm": 0.006331400014460087, + "learning_rate": 0.00019246098338158676, + "loss": 3.6906652450561523, + "step": 3075 + }, + { + "epoch": 42.72663755458515, + "grad_norm": 0.007149284239858389, + "learning_rate": 0.00019185531158027384, + "loss": 3.7055468559265137, + "step": 3076 + }, + { + "epoch": 42.74061135371179, + "grad_norm": 0.006395523902028799, + "learning_rate": 0.00019125070581273734, + "loss": 3.72355318069458, + "step": 3077 + }, + { + "epoch": 42.75458515283843, + "grad_norm": 0.007234807591885328, + "learning_rate": 0.0001906471665660978, + "loss": 3.6685009002685547, + "step": 3078 + }, + { + "epoch": 42.76855895196506, + "grad_norm": 0.006079623010009527, + "learning_rate": 0.00019004469432661478, + "loss": 3.7051684856414795, + "step": 3079 + }, + { + "epoch": 42.7825327510917, + "grad_norm": 0.00699756620451808, + "learning_rate": 0.0001894432895796907, + "loss": 3.6713948249816895, + "step": 3080 + }, + { + "epoch": 42.79650655021834, + "grad_norm": 0.006799036171287298, + "learning_rate": 0.00018884295280986531, + "loss": 3.7098617553710938, + "step": 3081 + }, + { + "epoch": 42.81048034934498, + "grad_norm": 0.006591275800019503, + "learning_rate": 0.0001882436845008197, + "loss": 3.685445785522461, + "step": 3082 + }, + { + "epoch": 42.82445414847162, + "grad_norm": 0.006709378212690353, + "learning_rate": 0.0001876454851353733, + "loss": 3.6773548126220703, + "step": 3083 + }, + { + "epoch": 42.838427947598255, + "grad_norm": 0.00651182746514678, + "learning_rate": 0.0001870483551954847, + "loss": 3.6529507637023926, + "step": 3084 + }, + { + "epoch": 42.852401746724894, + "grad_norm": 0.006494257599115372, + "learning_rate": 0.00018645229516225093, + "loss": 3.692331552505493, + "step": 3085 + }, + { + "epoch": 42.866375545851525, + "grad_norm": 0.006487370003014803, + "learning_rate": 0.00018585730551590619, + "loss": 3.679516315460205, + "step": 3086 + }, + { + "epoch": 42.880349344978164, + "grad_norm": 0.006413135677576065, + "learning_rate": 0.00018526338673582307, + "loss": 3.707998752593994, + "step": 3087 + }, + { + "epoch": 42.8943231441048, + "grad_norm": 0.006509249564260244, + "learning_rate": 0.00018467053930051156, + "loss": 3.7103686332702637, + "step": 3088 + }, + { + "epoch": 42.90829694323144, + "grad_norm": 0.006331913638859987, + "learning_rate": 0.0001840787636876177, + "loss": 3.6807055473327637, + "step": 3089 + }, + { + "epoch": 42.92227074235808, + "grad_norm": 0.006528888829052448, + "learning_rate": 0.00018348806037392473, + "loss": 3.7196922302246094, + "step": 3090 + }, + { + "epoch": 42.93624454148472, + "grad_norm": 0.006212258245795965, + "learning_rate": 0.0001828984298353515, + "loss": 3.716118335723877, + "step": 3091 + }, + { + "epoch": 42.95021834061135, + "grad_norm": 0.006641864310950041, + "learning_rate": 0.000182309872546953, + "loss": 3.6726081371307373, + "step": 3092 + }, + { + "epoch": 42.96419213973799, + "grad_norm": 0.006433268543332815, + "learning_rate": 0.0001817223889829186, + "loss": 3.690727710723877, + "step": 3093 + }, + { + "epoch": 42.97816593886463, + "grad_norm": 0.00644147302955389, + "learning_rate": 0.00018113597961657364, + "loss": 3.6998791694641113, + "step": 3094 + }, + { + "epoch": 42.992139737991266, + "grad_norm": 0.006477390881627798, + "learning_rate": 0.00018055064492037722, + "loss": 3.700925350189209, + "step": 3095 + }, + { + "epoch": 43.0, + "grad_norm": 0.0077974810265004635, + "learning_rate": 0.00017996638536592322, + "loss": 3.735599994659424, + "step": 3096 + }, + { + "epoch": 43.0, + "eval_loss": 4.376750946044922, + "eval_runtime": 52.3576, + "eval_samples_per_second": 46.641, + "eval_steps_per_second": 1.471, + "step": 3096 + }, + { + "epoch": 43.01397379912664, + "grad_norm": 0.006609050091356039, + "learning_rate": 0.00017938320142393897, + "loss": 3.6405727863311768, + "step": 3097 + }, + { + "epoch": 43.02794759825328, + "grad_norm": 0.006470015738159418, + "learning_rate": 0.00017880109356428522, + "loss": 3.6639411449432373, + "step": 3098 + }, + { + "epoch": 43.041921397379916, + "grad_norm": 0.006597487255930901, + "learning_rate": 0.00017822006225595596, + "loss": 3.6797537803649902, + "step": 3099 + }, + { + "epoch": 43.05589519650655, + "grad_norm": 0.0064637972973287106, + "learning_rate": 0.0001776401079670772, + "loss": 3.6798789501190186, + "step": 3100 + }, + { + "epoch": 43.069868995633186, + "grad_norm": 0.006872601341456175, + "learning_rate": 0.00017706123116490775, + "loss": 3.649874210357666, + "step": 3101 + }, + { + "epoch": 43.083842794759825, + "grad_norm": 0.0064028468914330006, + "learning_rate": 0.00017648343231583825, + "loss": 3.671036720275879, + "step": 3102 + }, + { + "epoch": 43.09781659388646, + "grad_norm": 0.006735685281455517, + "learning_rate": 0.00017590671188539084, + "loss": 3.6818368434906006, + "step": 3103 + }, + { + "epoch": 43.1117903930131, + "grad_norm": 0.006652928423136473, + "learning_rate": 0.00017533107033821895, + "loss": 3.660362720489502, + "step": 3104 + }, + { + "epoch": 43.12576419213974, + "grad_norm": 0.006564146373420954, + "learning_rate": 0.00017475650813810598, + "loss": 3.6444082260131836, + "step": 3105 + }, + { + "epoch": 43.13973799126637, + "grad_norm": 0.006476754322648048, + "learning_rate": 0.00017418302574796716, + "loss": 3.6836254596710205, + "step": 3106 + }, + { + "epoch": 43.15371179039301, + "grad_norm": 0.007108788937330246, + "learning_rate": 0.0001736106236298462, + "loss": 3.6424341201782227, + "step": 3107 + }, + { + "epoch": 43.16768558951965, + "grad_norm": 0.006570066325366497, + "learning_rate": 0.00017303930224491735, + "loss": 3.663379192352295, + "step": 3108 + }, + { + "epoch": 43.18165938864629, + "grad_norm": 0.0066614942625164986, + "learning_rate": 0.00017246906205348395, + "loss": 3.6918721199035645, + "step": 3109 + }, + { + "epoch": 43.19563318777293, + "grad_norm": 0.00704583153128624, + "learning_rate": 0.00017189990351497824, + "loss": 3.692559242248535, + "step": 3110 + }, + { + "epoch": 43.209606986899566, + "grad_norm": 0.006617981940507889, + "learning_rate": 0.00017133182708796126, + "loss": 3.6591439247131348, + "step": 3111 + }, + { + "epoch": 43.223580786026204, + "grad_norm": 0.006916946265846491, + "learning_rate": 0.00017076483323012127, + "loss": 3.7165699005126953, + "step": 3112 + }, + { + "epoch": 43.237554585152836, + "grad_norm": 0.006832561455667019, + "learning_rate": 0.00017019892239827534, + "loss": 3.7023603916168213, + "step": 3113 + }, + { + "epoch": 43.251528384279474, + "grad_norm": 0.00654558464884758, + "learning_rate": 0.00016963409504836777, + "loss": 3.6702237129211426, + "step": 3114 + }, + { + "epoch": 43.26550218340611, + "grad_norm": 0.007000794634222984, + "learning_rate": 0.0001690703516354698, + "loss": 3.69088077545166, + "step": 3115 + }, + { + "epoch": 43.27947598253275, + "grad_norm": 0.006473604589700699, + "learning_rate": 0.00016850769261377856, + "loss": 3.6641716957092285, + "step": 3116 + }, + { + "epoch": 43.29344978165939, + "grad_norm": 0.006787394639104605, + "learning_rate": 0.000167946118436619, + "loss": 3.6577839851379395, + "step": 3117 + }, + { + "epoch": 43.30742358078603, + "grad_norm": 0.006220053881406784, + "learning_rate": 0.00016738562955644144, + "loss": 3.6735315322875977, + "step": 3118 + }, + { + "epoch": 43.32139737991266, + "grad_norm": 0.006513087544590235, + "learning_rate": 0.00016682622642482113, + "loss": 3.624807119369507, + "step": 3119 + }, + { + "epoch": 43.3353711790393, + "grad_norm": 0.006552594713866711, + "learning_rate": 0.0001662679094924592, + "loss": 3.679218053817749, + "step": 3120 + }, + { + "epoch": 43.34934497816594, + "grad_norm": 0.006950687617063522, + "learning_rate": 0.0001657106792091814, + "loss": 3.6864023208618164, + "step": 3121 + }, + { + "epoch": 43.36331877729258, + "grad_norm": 0.006092227529734373, + "learning_rate": 0.0001651545360239387, + "loss": 3.693777084350586, + "step": 3122 + }, + { + "epoch": 43.377292576419215, + "grad_norm": 0.006924188230186701, + "learning_rate": 0.0001645994803848046, + "loss": 3.68534779548645, + "step": 3123 + }, + { + "epoch": 43.391266375545854, + "grad_norm": 0.0067315176129341125, + "learning_rate": 0.00016404551273897826, + "loss": 3.632850170135498, + "step": 3124 + }, + { + "epoch": 43.40524017467249, + "grad_norm": 0.007006707601249218, + "learning_rate": 0.00016349263353278146, + "loss": 3.6664111614227295, + "step": 3125 + }, + { + "epoch": 43.419213973799124, + "grad_norm": 0.0067031108774244785, + "learning_rate": 0.00016294084321165846, + "loss": 3.6863315105438232, + "step": 3126 + }, + { + "epoch": 43.43318777292576, + "grad_norm": 0.006502983160316944, + "learning_rate": 0.00016239014222017682, + "loss": 3.678041458129883, + "step": 3127 + }, + { + "epoch": 43.4471615720524, + "grad_norm": 0.006409413646906614, + "learning_rate": 0.00016184053100202665, + "loss": 3.692004919052124, + "step": 3128 + }, + { + "epoch": 43.46113537117904, + "grad_norm": 0.006251887418329716, + "learning_rate": 0.0001612920100000198, + "loss": 3.64532470703125, + "step": 3129 + }, + { + "epoch": 43.47510917030568, + "grad_norm": 0.0065657407976686954, + "learning_rate": 0.00016074457965608942, + "loss": 3.638624429702759, + "step": 3130 + }, + { + "epoch": 43.48908296943232, + "grad_norm": 0.006246988195925951, + "learning_rate": 0.00016019824041129004, + "loss": 3.668254852294922, + "step": 3131 + }, + { + "epoch": 43.50305676855895, + "grad_norm": 0.0065050385892391205, + "learning_rate": 0.0001596529927057984, + "loss": 3.6901469230651855, + "step": 3132 + }, + { + "epoch": 43.51703056768559, + "grad_norm": 0.006240491755306721, + "learning_rate": 0.00015910883697890977, + "loss": 3.6796071529388428, + "step": 3133 + }, + { + "epoch": 43.531004366812226, + "grad_norm": 0.006674987263977528, + "learning_rate": 0.00015856577366904075, + "loss": 3.7001953125, + "step": 3134 + }, + { + "epoch": 43.544978165938865, + "grad_norm": 0.006260441616177559, + "learning_rate": 0.00015802380321372776, + "loss": 3.6646556854248047, + "step": 3135 + }, + { + "epoch": 43.5589519650655, + "grad_norm": 0.006696469150483608, + "learning_rate": 0.00015748292604962687, + "loss": 3.6791043281555176, + "step": 3136 + }, + { + "epoch": 43.57292576419214, + "grad_norm": 0.006716041825711727, + "learning_rate": 0.00015694314261251245, + "loss": 3.710442066192627, + "step": 3137 + }, + { + "epoch": 43.58689956331878, + "grad_norm": 0.006856406573206186, + "learning_rate": 0.0001564044533372785, + "loss": 3.6615681648254395, + "step": 3138 + }, + { + "epoch": 43.60087336244541, + "grad_norm": 0.0064330073073506355, + "learning_rate": 0.00015586685865793748, + "loss": 3.707085609436035, + "step": 3139 + }, + { + "epoch": 43.61484716157205, + "grad_norm": 0.006748909596353769, + "learning_rate": 0.00015533035900761955, + "loss": 3.6814656257629395, + "step": 3140 + }, + { + "epoch": 43.62882096069869, + "grad_norm": 0.006306119728833437, + "learning_rate": 0.00015479495481857282, + "loss": 3.6595420837402344, + "step": 3141 + }, + { + "epoch": 43.64279475982533, + "grad_norm": 0.0061928280629217625, + "learning_rate": 0.00015426064652216255, + "loss": 3.686880111694336, + "step": 3142 + }, + { + "epoch": 43.65676855895197, + "grad_norm": 0.006499494891613722, + "learning_rate": 0.00015372743454887173, + "loss": 3.6901583671569824, + "step": 3143 + }, + { + "epoch": 43.670742358078606, + "grad_norm": 0.005914116278290749, + "learning_rate": 0.00015319531932829917, + "loss": 3.699162006378174, + "step": 3144 + }, + { + "epoch": 43.68471615720524, + "grad_norm": 0.006499269045889378, + "learning_rate": 0.00015266430128916075, + "loss": 3.6740684509277344, + "step": 3145 + }, + { + "epoch": 43.698689956331876, + "grad_norm": 0.006433961447328329, + "learning_rate": 0.000152134380859288, + "loss": 3.684006929397583, + "step": 3146 + }, + { + "epoch": 43.712663755458514, + "grad_norm": 0.006057603284716606, + "learning_rate": 0.00015160555846562826, + "loss": 3.6612191200256348, + "step": 3147 + }, + { + "epoch": 43.72663755458515, + "grad_norm": 0.006337927188724279, + "learning_rate": 0.00015107783453424456, + "loss": 3.6981983184814453, + "step": 3148 + }, + { + "epoch": 43.74061135371179, + "grad_norm": 0.006348243448883295, + "learning_rate": 0.00015055120949031365, + "loss": 3.6667895317077637, + "step": 3149 + }, + { + "epoch": 43.75458515283843, + "grad_norm": 0.0060755726881325245, + "learning_rate": 0.00015002568375812876, + "loss": 3.6844797134399414, + "step": 3150 + }, + { + "epoch": 43.76855895196506, + "grad_norm": 0.006059548351913691, + "learning_rate": 0.0001495012577610959, + "loss": 3.6676461696624756, + "step": 3151 + }, + { + "epoch": 43.7825327510917, + "grad_norm": 0.006432189140468836, + "learning_rate": 0.00014897793192173602, + "loss": 3.7036523818969727, + "step": 3152 + }, + { + "epoch": 43.79650655021834, + "grad_norm": 0.006043825298547745, + "learning_rate": 0.00014845570666168249, + "loss": 3.7043371200561523, + "step": 3153 + }, + { + "epoch": 43.81048034934498, + "grad_norm": 0.0066383229568600655, + "learning_rate": 0.0001479345824016837, + "loss": 3.6830334663391113, + "step": 3154 + }, + { + "epoch": 43.82445414847162, + "grad_norm": 0.006384526379406452, + "learning_rate": 0.00014741455956160005, + "loss": 3.6746675968170166, + "step": 3155 + }, + { + "epoch": 43.838427947598255, + "grad_norm": 0.006477308925241232, + "learning_rate": 0.0001468956385604038, + "loss": 3.6788225173950195, + "step": 3156 + }, + { + "epoch": 43.852401746724894, + "grad_norm": 0.006130026653409004, + "learning_rate": 0.00014637781981618054, + "loss": 3.6821796894073486, + "step": 3157 + }, + { + "epoch": 43.866375545851525, + "grad_norm": 0.006257957778871059, + "learning_rate": 0.00014586110374612777, + "loss": 3.683966636657715, + "step": 3158 + }, + { + "epoch": 43.880349344978164, + "grad_norm": 0.006265794392675161, + "learning_rate": 0.00014534549076655425, + "loss": 3.691389560699463, + "step": 3159 + }, + { + "epoch": 43.8943231441048, + "grad_norm": 0.006294097751379013, + "learning_rate": 0.0001448309812928794, + "loss": 3.659594774246216, + "step": 3160 + }, + { + "epoch": 43.90829694323144, + "grad_norm": 0.006432943977415562, + "learning_rate": 0.00014431757573963504, + "loss": 3.665633201599121, + "step": 3161 + }, + { + "epoch": 43.92227074235808, + "grad_norm": 0.006102072540670633, + "learning_rate": 0.0001438052745204626, + "loss": 3.6348228454589844, + "step": 3162 + }, + { + "epoch": 43.93624454148472, + "grad_norm": 0.00636800192296505, + "learning_rate": 0.00014329407804811333, + "loss": 3.624040365219116, + "step": 3163 + }, + { + "epoch": 43.95021834061135, + "grad_norm": 0.006060190033167601, + "learning_rate": 0.00014278398673444966, + "loss": 3.6986045837402344, + "step": 3164 + }, + { + "epoch": 43.96419213973799, + "grad_norm": 0.0061101666651666164, + "learning_rate": 0.00014227500099044254, + "loss": 3.6816675662994385, + "step": 3165 + }, + { + "epoch": 43.97816593886463, + "grad_norm": 0.006048897281289101, + "learning_rate": 0.00014176712122617293, + "loss": 3.6739861965179443, + "step": 3166 + }, + { + "epoch": 43.992139737991266, + "grad_norm": 0.005954945459961891, + "learning_rate": 0.00014126034785082997, + "loss": 3.679354190826416, + "step": 3167 + }, + { + "epoch": 44.0, + "grad_norm": 0.007545145694166422, + "learning_rate": 0.0001407546812727119, + "loss": 3.7448782920837402, + "step": 3168 + }, + { + "epoch": 44.0, + "eval_loss": 4.377684116363525, + "eval_runtime": 51.0591, + "eval_samples_per_second": 47.827, + "eval_steps_per_second": 1.508, + "step": 3168 + }, + { + "epoch": 44.01397379912664, + "grad_norm": 0.00643246341496706, + "learning_rate": 0.00014025012189922562, + "loss": 3.6544783115386963, + "step": 3169 + }, + { + "epoch": 44.02794759825328, + "grad_norm": 0.006309461314231157, + "learning_rate": 0.00013974667013688495, + "loss": 3.701963186264038, + "step": 3170 + }, + { + "epoch": 44.041921397379916, + "grad_norm": 0.006177262403070927, + "learning_rate": 0.00013924432639131215, + "loss": 3.649641275405884, + "step": 3171 + }, + { + "epoch": 44.05589519650655, + "grad_norm": 0.006319771986454725, + "learning_rate": 0.00013874309106723636, + "loss": 3.67417573928833, + "step": 3172 + }, + { + "epoch": 44.069868995633186, + "grad_norm": 0.006548433564603329, + "learning_rate": 0.00013824296456849426, + "loss": 3.6257686614990234, + "step": 3173 + }, + { + "epoch": 44.083842794759825, + "grad_norm": 0.006148583721369505, + "learning_rate": 0.00013774394729802804, + "loss": 3.675339460372925, + "step": 3174 + }, + { + "epoch": 44.09781659388646, + "grad_norm": 0.0063438634388148785, + "learning_rate": 0.00013724603965788716, + "loss": 3.666316270828247, + "step": 3175 + }, + { + "epoch": 44.1117903930131, + "grad_norm": 0.006417148746550083, + "learning_rate": 0.00013674924204922732, + "loss": 3.6692237854003906, + "step": 3176 + }, + { + "epoch": 44.12576419213974, + "grad_norm": 0.0062498473562300205, + "learning_rate": 0.00013625355487230872, + "loss": 3.70363187789917, + "step": 3177 + }, + { + "epoch": 44.13973799126637, + "grad_norm": 0.006249656435102224, + "learning_rate": 0.00013575897852649796, + "loss": 3.691108226776123, + "step": 3178 + }, + { + "epoch": 44.15371179039301, + "grad_norm": 0.006240040063858032, + "learning_rate": 0.0001352655134102657, + "loss": 3.6781342029571533, + "step": 3179 + }, + { + "epoch": 44.16768558951965, + "grad_norm": 0.006186297629028559, + "learning_rate": 0.00013477315992118885, + "loss": 3.644334316253662, + "step": 3180 + }, + { + "epoch": 44.18165938864629, + "grad_norm": 0.006283515132963657, + "learning_rate": 0.00013428191845594675, + "loss": 3.6691460609436035, + "step": 3181 + }, + { + "epoch": 44.19563318777293, + "grad_norm": 0.0064705004915595055, + "learning_rate": 0.00013379178941032427, + "loss": 3.64168119430542, + "step": 3182 + }, + { + "epoch": 44.209606986899566, + "grad_norm": 0.006210929248481989, + "learning_rate": 0.00013330277317920959, + "loss": 3.66428804397583, + "step": 3183 + }, + { + "epoch": 44.223580786026204, + "grad_norm": 0.006095333956182003, + "learning_rate": 0.00013281487015659395, + "loss": 3.6205215454101562, + "step": 3184 + }, + { + "epoch": 44.237554585152836, + "grad_norm": 0.006364993751049042, + "learning_rate": 0.00013232808073557244, + "loss": 3.662804126739502, + "step": 3185 + }, + { + "epoch": 44.251528384279474, + "grad_norm": 0.006166066508740187, + "learning_rate": 0.00013184240530834173, + "loss": 3.702338218688965, + "step": 3186 + }, + { + "epoch": 44.26550218340611, + "grad_norm": 0.006147229578346014, + "learning_rate": 0.00013135784426620282, + "loss": 3.660149097442627, + "step": 3187 + }, + { + "epoch": 44.27947598253275, + "grad_norm": 0.006086066365242004, + "learning_rate": 0.00013087439799955674, + "loss": 3.652472972869873, + "step": 3188 + }, + { + "epoch": 44.29344978165939, + "grad_norm": 0.006371276918798685, + "learning_rate": 0.00013039206689790815, + "loss": 3.6471242904663086, + "step": 3189 + }, + { + "epoch": 44.30742358078603, + "grad_norm": 0.005952878389507532, + "learning_rate": 0.000129910851349862, + "loss": 3.6450729370117188, + "step": 3190 + }, + { + "epoch": 44.32139737991266, + "grad_norm": 0.006385558284819126, + "learning_rate": 0.00012943075174312536, + "loss": 3.6730473041534424, + "step": 3191 + }, + { + "epoch": 44.3353711790393, + "grad_norm": 0.005896324757486582, + "learning_rate": 0.00012895176846450582, + "loss": 3.6646409034729004, + "step": 3192 + }, + { + "epoch": 44.34934497816594, + "grad_norm": 0.006053723394870758, + "learning_rate": 0.00012847390189991121, + "loss": 3.6568682193756104, + "step": 3193 + }, + { + "epoch": 44.36331877729258, + "grad_norm": 0.005929332692176104, + "learning_rate": 0.00012799715243435027, + "loss": 3.7030997276306152, + "step": 3194 + }, + { + "epoch": 44.377292576419215, + "grad_norm": 0.006153058726340532, + "learning_rate": 0.00012752152045193116, + "loss": 3.6949849128723145, + "step": 3195 + }, + { + "epoch": 44.391266375545854, + "grad_norm": 0.006193236447870731, + "learning_rate": 0.0001270470063358624, + "loss": 3.6040501594543457, + "step": 3196 + }, + { + "epoch": 44.40524017467249, + "grad_norm": 0.005849961191415787, + "learning_rate": 0.0001265736104684509, + "loss": 3.602717399597168, + "step": 3197 + }, + { + "epoch": 44.419213973799124, + "grad_norm": 0.0060405549593269825, + "learning_rate": 0.00012610133323110356, + "loss": 3.668717861175537, + "step": 3198 + }, + { + "epoch": 44.43318777292576, + "grad_norm": 0.0062002381309866905, + "learning_rate": 0.00012563017500432565, + "loss": 3.6333389282226562, + "step": 3199 + }, + { + "epoch": 44.4471615720524, + "grad_norm": 0.006093314848840237, + "learning_rate": 0.00012516013616772065, + "loss": 3.662696361541748, + "step": 3200 + }, + { + "epoch": 44.46113537117904, + "grad_norm": 0.005765238776803017, + "learning_rate": 0.00012469121709999026, + "loss": 3.6843791007995605, + "step": 3201 + }, + { + "epoch": 44.47510917030568, + "grad_norm": 0.005989165045320988, + "learning_rate": 0.00012422341817893448, + "loss": 3.6904735565185547, + "step": 3202 + }, + { + "epoch": 44.48908296943232, + "grad_norm": 0.006150389555841684, + "learning_rate": 0.00012375673978145022, + "loss": 3.6614508628845215, + "step": 3203 + }, + { + "epoch": 44.50305676855895, + "grad_norm": 0.005898299627006054, + "learning_rate": 0.00012329118228353178, + "loss": 3.7193222045898438, + "step": 3204 + }, + { + "epoch": 44.51703056768559, + "grad_norm": 0.0058402069844305515, + "learning_rate": 0.00012282674606027016, + "loss": 3.669564723968506, + "step": 3205 + }, + { + "epoch": 44.531004366812226, + "grad_norm": 0.006015103310346603, + "learning_rate": 0.00012236343148585413, + "loss": 3.686734914779663, + "step": 3206 + }, + { + "epoch": 44.544978165938865, + "grad_norm": 0.005990280769765377, + "learning_rate": 0.00012190123893356703, + "loss": 3.664945125579834, + "step": 3207 + }, + { + "epoch": 44.5589519650655, + "grad_norm": 0.006110248621553183, + "learning_rate": 0.00012144016877578942, + "loss": 3.6546223163604736, + "step": 3208 + }, + { + "epoch": 44.57292576419214, + "grad_norm": 0.006006988696753979, + "learning_rate": 0.000120980221383997, + "loss": 3.669175386428833, + "step": 3209 + }, + { + "epoch": 44.58689956331878, + "grad_norm": 0.006008700001984835, + "learning_rate": 0.00012052139712876141, + "loss": 3.672668695449829, + "step": 3210 + }, + { + "epoch": 44.60087336244541, + "grad_norm": 0.005862198304384947, + "learning_rate": 0.00012006369637974875, + "loss": 3.6824991703033447, + "step": 3211 + }, + { + "epoch": 44.61484716157205, + "grad_norm": 0.005960221868008375, + "learning_rate": 0.00011960711950572005, + "loss": 3.6258697509765625, + "step": 3212 + }, + { + "epoch": 44.62882096069869, + "grad_norm": 0.005813480354845524, + "learning_rate": 0.00011915166687453174, + "loss": 3.6463029384613037, + "step": 3213 + }, + { + "epoch": 44.64279475982533, + "grad_norm": 0.006002446636557579, + "learning_rate": 0.0001186973388531332, + "loss": 3.721970558166504, + "step": 3214 + }, + { + "epoch": 44.65676855895197, + "grad_norm": 0.006029803771525621, + "learning_rate": 0.00011824413580756844, + "loss": 3.685335159301758, + "step": 3215 + }, + { + "epoch": 44.670742358078606, + "grad_norm": 0.005933200940489769, + "learning_rate": 0.0001177920581029752, + "loss": 3.6834492683410645, + "step": 3216 + }, + { + "epoch": 44.68471615720524, + "grad_norm": 0.005942870397120714, + "learning_rate": 0.00011734110610358423, + "loss": 3.670149326324463, + "step": 3217 + }, + { + "epoch": 44.698689956331876, + "grad_norm": 0.005972386337816715, + "learning_rate": 0.00011689128017271901, + "loss": 3.6798110008239746, + "step": 3218 + }, + { + "epoch": 44.712663755458514, + "grad_norm": 0.005964178126305342, + "learning_rate": 0.00011644258067279662, + "loss": 3.6620044708251953, + "step": 3219 + }, + { + "epoch": 44.72663755458515, + "grad_norm": 0.0059594810009002686, + "learning_rate": 0.00011599500796532605, + "loss": 3.683058738708496, + "step": 3220 + }, + { + "epoch": 44.74061135371179, + "grad_norm": 0.005901646800339222, + "learning_rate": 0.00011554856241090844, + "loss": 3.6791443824768066, + "step": 3221 + }, + { + "epoch": 44.75458515283843, + "grad_norm": 0.006178564857691526, + "learning_rate": 0.00011510324436923708, + "loss": 3.6620421409606934, + "step": 3222 + }, + { + "epoch": 44.76855895196506, + "grad_norm": 0.005973632447421551, + "learning_rate": 0.0001146590541990962, + "loss": 3.6821961402893066, + "step": 3223 + }, + { + "epoch": 44.7825327510917, + "grad_norm": 0.005914160516113043, + "learning_rate": 0.00011421599225836265, + "loss": 3.6588668823242188, + "step": 3224 + }, + { + "epoch": 44.79650655021834, + "grad_norm": 0.005887249484658241, + "learning_rate": 0.00011377405890400266, + "loss": 3.683187484741211, + "step": 3225 + }, + { + "epoch": 44.81048034934498, + "grad_norm": 0.006068149581551552, + "learning_rate": 0.00011333325449207415, + "loss": 3.6633477210998535, + "step": 3226 + }, + { + "epoch": 44.82445414847162, + "grad_norm": 0.005748802796006203, + "learning_rate": 0.00011289357937772547, + "loss": 3.676607608795166, + "step": 3227 + }, + { + "epoch": 44.838427947598255, + "grad_norm": 0.006032158620655537, + "learning_rate": 0.00011245503391519489, + "loss": 3.652369499206543, + "step": 3228 + }, + { + "epoch": 44.852401746724894, + "grad_norm": 0.00615714630112052, + "learning_rate": 0.00011201761845781049, + "loss": 3.7210354804992676, + "step": 3229 + }, + { + "epoch": 44.866375545851525, + "grad_norm": 0.0059432960115373135, + "learning_rate": 0.00011158133335799003, + "loss": 3.640336036682129, + "step": 3230 + }, + { + "epoch": 44.880349344978164, + "grad_norm": 0.0058724163100123405, + "learning_rate": 0.00011114617896724026, + "loss": 3.638444185256958, + "step": 3231 + }, + { + "epoch": 44.8943231441048, + "grad_norm": 0.006073266267776489, + "learning_rate": 0.00011071215563615752, + "loss": 3.680121421813965, + "step": 3232 + }, + { + "epoch": 44.90829694323144, + "grad_norm": 0.005904203746467829, + "learning_rate": 0.00011027926371442662, + "loss": 3.6883950233459473, + "step": 3233 + }, + { + "epoch": 44.92227074235808, + "grad_norm": 0.005889971274882555, + "learning_rate": 0.00010984750355082045, + "loss": 3.6875810623168945, + "step": 3234 + }, + { + "epoch": 44.93624454148472, + "grad_norm": 0.005878197029232979, + "learning_rate": 0.00010941687549320068, + "loss": 3.6341497898101807, + "step": 3235 + }, + { + "epoch": 44.95021834061135, + "grad_norm": 0.005862570833414793, + "learning_rate": 0.00010898737988851677, + "loss": 3.648092746734619, + "step": 3236 + }, + { + "epoch": 44.96419213973799, + "grad_norm": 0.0058378661051392555, + "learning_rate": 0.00010855901708280482, + "loss": 3.627786636352539, + "step": 3237 + }, + { + "epoch": 44.97816593886463, + "grad_norm": 0.0057954080402851105, + "learning_rate": 0.0001081317874211895, + "loss": 3.636265754699707, + "step": 3238 + }, + { + "epoch": 44.992139737991266, + "grad_norm": 0.005763230379670858, + "learning_rate": 0.00010770569124788201, + "loss": 3.6991324424743652, + "step": 3239 + }, + { + "epoch": 45.0, + "grad_norm": 0.007435750216245651, + "learning_rate": 0.00010728072890618029, + "loss": 3.698052406311035, + "step": 3240 + }, + { + "epoch": 45.0, + "eval_loss": 4.379567623138428, + "eval_runtime": 50.7833, + "eval_samples_per_second": 48.087, + "eval_steps_per_second": 1.516, + "step": 3240 + }, + { + "epoch": 45.01397379912664, + "grad_norm": 0.0059455581940710545, + "learning_rate": 0.00010685690073846878, + "loss": 3.649308204650879, + "step": 3241 + }, + { + "epoch": 45.02794759825328, + "grad_norm": 0.006021182518452406, + "learning_rate": 0.00010643420708621834, + "loss": 3.6339311599731445, + "step": 3242 + }, + { + "epoch": 45.041921397379916, + "grad_norm": 0.0061815716326236725, + "learning_rate": 0.00010601264828998569, + "loss": 3.658045530319214, + "step": 3243 + }, + { + "epoch": 45.05589519650655, + "grad_norm": 0.005834038369357586, + "learning_rate": 0.0001055922246894127, + "loss": 3.6551108360290527, + "step": 3244 + }, + { + "epoch": 45.069868995633186, + "grad_norm": 0.005903543904423714, + "learning_rate": 0.0001051729366232273, + "loss": 3.6415858268737793, + "step": 3245 + }, + { + "epoch": 45.083842794759825, + "grad_norm": 0.006033484358340502, + "learning_rate": 0.00010475478442924231, + "loss": 3.607917308807373, + "step": 3246 + }, + { + "epoch": 45.09781659388646, + "grad_norm": 0.005784410983324051, + "learning_rate": 0.00010433776844435538, + "loss": 3.641299247741699, + "step": 3247 + }, + { + "epoch": 45.1117903930131, + "grad_norm": 0.006221757270395756, + "learning_rate": 0.00010392188900454902, + "loss": 3.632112979888916, + "step": 3248 + }, + { + "epoch": 45.12576419213974, + "grad_norm": 0.005921157542616129, + "learning_rate": 0.00010350714644488925, + "loss": 3.6394639015197754, + "step": 3249 + }, + { + "epoch": 45.13973799126637, + "grad_norm": 0.005843059625476599, + "learning_rate": 0.00010309354109952728, + "loss": 3.6688497066497803, + "step": 3250 + }, + { + "epoch": 45.15371179039301, + "grad_norm": 0.006184449419379234, + "learning_rate": 0.000102681073301697, + "loss": 3.62900972366333, + "step": 3251 + }, + { + "epoch": 45.16768558951965, + "grad_norm": 0.005910454783588648, + "learning_rate": 0.00010226974338371635, + "loss": 3.639268398284912, + "step": 3252 + }, + { + "epoch": 45.18165938864629, + "grad_norm": 0.006613037548959255, + "learning_rate": 0.00010185955167698657, + "loss": 3.651017665863037, + "step": 3253 + }, + { + "epoch": 45.19563318777293, + "grad_norm": 0.006017698906362057, + "learning_rate": 0.00010145049851199155, + "loss": 3.6830077171325684, + "step": 3254 + }, + { + "epoch": 45.209606986899566, + "grad_norm": 0.006236966699361801, + "learning_rate": 0.0001010425842182984, + "loss": 3.6856958866119385, + "step": 3255 + }, + { + "epoch": 45.223580786026204, + "grad_norm": 0.0062433164566755295, + "learning_rate": 0.0001006358091245557, + "loss": 3.7161612510681152, + "step": 3256 + }, + { + "epoch": 45.237554585152836, + "grad_norm": 0.006063258275389671, + "learning_rate": 0.00010023017355849525, + "loss": 3.649462938308716, + "step": 3257 + }, + { + "epoch": 45.251528384279474, + "grad_norm": 0.006399665027856827, + "learning_rate": 9.982567784693017e-05, + "loss": 3.624393939971924, + "step": 3258 + }, + { + "epoch": 45.26550218340611, + "grad_norm": 0.006117025390267372, + "learning_rate": 9.94223223157554e-05, + "loss": 3.6675968170166016, + "step": 3259 + }, + { + "epoch": 45.27947598253275, + "grad_norm": 0.006158605683594942, + "learning_rate": 9.902010728994665e-05, + "loss": 3.6349360942840576, + "step": 3260 + }, + { + "epoch": 45.29344978165939, + "grad_norm": 0.0061307125724852085, + "learning_rate": 9.861903309356244e-05, + "loss": 3.6786162853240967, + "step": 3261 + }, + { + "epoch": 45.30742358078603, + "grad_norm": 0.005947567522525787, + "learning_rate": 9.821910004973994e-05, + "loss": 3.651076316833496, + "step": 3262 + }, + { + "epoch": 45.32139737991266, + "grad_norm": 0.0059933350421488285, + "learning_rate": 9.782030848069845e-05, + "loss": 3.6713902950286865, + "step": 3263 + }, + { + "epoch": 45.3353711790393, + "grad_norm": 0.006069707218557596, + "learning_rate": 9.74226587077372e-05, + "loss": 3.6770577430725098, + "step": 3264 + }, + { + "epoch": 45.34934497816594, + "grad_norm": 0.005641316995024681, + "learning_rate": 9.702615105123522e-05, + "loss": 3.6663057804107666, + "step": 3265 + }, + { + "epoch": 45.36331877729258, + "grad_norm": 0.006138736382126808, + "learning_rate": 9.663078583065183e-05, + "loss": 3.7058024406433105, + "step": 3266 + }, + { + "epoch": 45.377292576419215, + "grad_norm": 0.006074203178286552, + "learning_rate": 9.623656336452509e-05, + "loss": 3.684798002243042, + "step": 3267 + }, + { + "epoch": 45.391266375545854, + "grad_norm": 0.006269104778766632, + "learning_rate": 9.58434839704738e-05, + "loss": 3.653730630874634, + "step": 3268 + }, + { + "epoch": 45.40524017467249, + "grad_norm": 0.005954077932983637, + "learning_rate": 9.545154796519407e-05, + "loss": 3.631556272506714, + "step": 3269 + }, + { + "epoch": 45.419213973799124, + "grad_norm": 0.006305301561951637, + "learning_rate": 9.506075566446213e-05, + "loss": 3.652346134185791, + "step": 3270 + }, + { + "epoch": 45.43318777292576, + "grad_norm": 0.005923055112361908, + "learning_rate": 9.46711073831322e-05, + "loss": 3.644423007965088, + "step": 3271 + }, + { + "epoch": 45.4471615720524, + "grad_norm": 0.005840525031089783, + "learning_rate": 9.428260343513663e-05, + "loss": 3.664320468902588, + "step": 3272 + }, + { + "epoch": 45.46113537117904, + "grad_norm": 0.0058747464790940285, + "learning_rate": 9.389524413348661e-05, + "loss": 3.693209409713745, + "step": 3273 + }, + { + "epoch": 45.47510917030568, + "grad_norm": 0.005935903172940016, + "learning_rate": 9.350902979027003e-05, + "loss": 3.655853509902954, + "step": 3274 + }, + { + "epoch": 45.48908296943232, + "grad_norm": 0.005829546134918928, + "learning_rate": 9.312396071665273e-05, + "loss": 3.645937442779541, + "step": 3275 + }, + { + "epoch": 45.50305676855895, + "grad_norm": 0.005715377628803253, + "learning_rate": 9.274003722287829e-05, + "loss": 3.660888671875, + "step": 3276 + }, + { + "epoch": 45.51703056768559, + "grad_norm": 0.005914001259952784, + "learning_rate": 9.235725961826662e-05, + "loss": 3.6928553581237793, + "step": 3277 + }, + { + "epoch": 45.531004366812226, + "grad_norm": 0.006055662874132395, + "learning_rate": 9.197562821121495e-05, + "loss": 3.6630935668945312, + "step": 3278 + }, + { + "epoch": 45.544978165938865, + "grad_norm": 0.005942841526120901, + "learning_rate": 9.159514330919671e-05, + "loss": 3.6434600353240967, + "step": 3279 + }, + { + "epoch": 45.5589519650655, + "grad_norm": 0.005879723932594061, + "learning_rate": 9.121580521876192e-05, + "loss": 3.6828930377960205, + "step": 3280 + }, + { + "epoch": 45.57292576419214, + "grad_norm": 0.005810475442558527, + "learning_rate": 9.083761424553615e-05, + "loss": 3.6644086837768555, + "step": 3281 + }, + { + "epoch": 45.58689956331878, + "grad_norm": 0.0058500380255281925, + "learning_rate": 9.046057069422091e-05, + "loss": 3.6613426208496094, + "step": 3282 + }, + { + "epoch": 45.60087336244541, + "grad_norm": 0.005881919525563717, + "learning_rate": 9.008467486859382e-05, + "loss": 3.646371841430664, + "step": 3283 + }, + { + "epoch": 45.61484716157205, + "grad_norm": 0.00608665868639946, + "learning_rate": 8.970992707150702e-05, + "loss": 3.6912505626678467, + "step": 3284 + }, + { + "epoch": 45.62882096069869, + "grad_norm": 0.0059072356671094894, + "learning_rate": 8.933632760488831e-05, + "loss": 3.6508445739746094, + "step": 3285 + }, + { + "epoch": 45.64279475982533, + "grad_norm": 0.00589236943051219, + "learning_rate": 8.896387676973956e-05, + "loss": 3.636539936065674, + "step": 3286 + }, + { + "epoch": 45.65676855895197, + "grad_norm": 0.006131782662123442, + "learning_rate": 8.859257486613818e-05, + "loss": 3.643298625946045, + "step": 3287 + }, + { + "epoch": 45.670742358078606, + "grad_norm": 0.006039046682417393, + "learning_rate": 8.82224221932353e-05, + "loss": 3.6826415061950684, + "step": 3288 + }, + { + "epoch": 45.68471615720524, + "grad_norm": 0.005942283198237419, + "learning_rate": 8.785341904925602e-05, + "loss": 3.6603708267211914, + "step": 3289 + }, + { + "epoch": 45.698689956331876, + "grad_norm": 0.006028817500919104, + "learning_rate": 8.748556573149962e-05, + "loss": 3.6836555004119873, + "step": 3290 + }, + { + "epoch": 45.712663755458514, + "grad_norm": 0.006245827302336693, + "learning_rate": 8.711886253633896e-05, + "loss": 3.689937114715576, + "step": 3291 + }, + { + "epoch": 45.72663755458515, + "grad_norm": 0.005829308647662401, + "learning_rate": 8.675330975922034e-05, + "loss": 3.646091938018799, + "step": 3292 + }, + { + "epoch": 45.74061135371179, + "grad_norm": 0.005799743812531233, + "learning_rate": 8.63889076946624e-05, + "loss": 3.667741298675537, + "step": 3293 + }, + { + "epoch": 45.75458515283843, + "grad_norm": 0.006182873621582985, + "learning_rate": 8.602565663625812e-05, + "loss": 3.6704702377319336, + "step": 3294 + }, + { + "epoch": 45.76855895196506, + "grad_norm": 0.0059661525301635265, + "learning_rate": 8.566355687667166e-05, + "loss": 3.6082215309143066, + "step": 3295 + }, + { + "epoch": 45.7825327510917, + "grad_norm": 0.006096175871789455, + "learning_rate": 8.530260870764065e-05, + "loss": 3.6541519165039062, + "step": 3296 + }, + { + "epoch": 45.79650655021834, + "grad_norm": 0.005957531277090311, + "learning_rate": 8.494281241997411e-05, + "loss": 3.6488404273986816, + "step": 3297 + }, + { + "epoch": 45.81048034934498, + "grad_norm": 0.006362909451127052, + "learning_rate": 8.458416830355393e-05, + "loss": 3.6703639030456543, + "step": 3298 + }, + { + "epoch": 45.82445414847162, + "grad_norm": 0.005813185591250658, + "learning_rate": 8.422667664733304e-05, + "loss": 3.6723361015319824, + "step": 3299 + }, + { + "epoch": 45.838427947598255, + "grad_norm": 0.005782100837677717, + "learning_rate": 8.387033773933557e-05, + "loss": 3.63480806350708, + "step": 3300 + }, + { + "epoch": 45.852401746724894, + "grad_norm": 0.006054454483091831, + "learning_rate": 8.351515186665794e-05, + "loss": 3.6608922481536865, + "step": 3301 + }, + { + "epoch": 45.866375545851525, + "grad_norm": 0.006052988115698099, + "learning_rate": 8.316111931546666e-05, + "loss": 3.6687750816345215, + "step": 3302 + }, + { + "epoch": 45.880349344978164, + "grad_norm": 0.006084649357944727, + "learning_rate": 8.280824037099941e-05, + "loss": 3.6396217346191406, + "step": 3303 + }, + { + "epoch": 45.8943231441048, + "grad_norm": 0.005772392265498638, + "learning_rate": 8.245651531756433e-05, + "loss": 3.7278900146484375, + "step": 3304 + }, + { + "epoch": 45.90829694323144, + "grad_norm": 0.005887683946639299, + "learning_rate": 8.210594443854007e-05, + "loss": 3.6623833179473877, + "step": 3305 + }, + { + "epoch": 45.92227074235808, + "grad_norm": 0.005802373867481947, + "learning_rate": 8.175652801637535e-05, + "loss": 3.680635690689087, + "step": 3306 + }, + { + "epoch": 45.93624454148472, + "grad_norm": 0.005748553667217493, + "learning_rate": 8.14082663325883e-05, + "loss": 3.6317763328552246, + "step": 3307 + }, + { + "epoch": 45.95021834061135, + "grad_norm": 0.005717726889997721, + "learning_rate": 8.106115966776743e-05, + "loss": 3.6421971321105957, + "step": 3308 + }, + { + "epoch": 45.96419213973799, + "grad_norm": 0.005919075571000576, + "learning_rate": 8.071520830156999e-05, + "loss": 3.726754665374756, + "step": 3309 + }, + { + "epoch": 45.97816593886463, + "grad_norm": 0.005714495666325092, + "learning_rate": 8.037041251272334e-05, + "loss": 3.637263298034668, + "step": 3310 + }, + { + "epoch": 45.992139737991266, + "grad_norm": 0.005621272139251232, + "learning_rate": 8.002677257902252e-05, + "loss": 3.6489784717559814, + "step": 3311 + }, + { + "epoch": 46.0, + "grad_norm": 0.007174668367952108, + "learning_rate": 7.968428877733201e-05, + "loss": 3.67305326461792, + "step": 3312 + }, + { + "epoch": 46.0, + "eval_loss": 4.380944728851318, + "eval_runtime": 51.6456, + "eval_samples_per_second": 47.284, + "eval_steps_per_second": 1.491, + "step": 3312 + }, + { + "epoch": 46.01397379912664, + "grad_norm": 0.00578242726624012, + "learning_rate": 7.934296138358568e-05, + "loss": 3.6154122352600098, + "step": 3313 + }, + { + "epoch": 46.02794759825328, + "grad_norm": 0.005650710314512253, + "learning_rate": 7.900279067278403e-05, + "loss": 3.6636948585510254, + "step": 3314 + }, + { + "epoch": 46.041921397379916, + "grad_norm": 0.0056931874714791775, + "learning_rate": 7.866377691899657e-05, + "loss": 3.636855125427246, + "step": 3315 + }, + { + "epoch": 46.05589519650655, + "grad_norm": 0.005795546341687441, + "learning_rate": 7.832592039536082e-05, + "loss": 3.6582765579223633, + "step": 3316 + }, + { + "epoch": 46.069868995633186, + "grad_norm": 0.005859258119016886, + "learning_rate": 7.798922137408176e-05, + "loss": 3.6686697006225586, + "step": 3317 + }, + { + "epoch": 46.083842794759825, + "grad_norm": 0.005823366343975067, + "learning_rate": 7.765368012643131e-05, + "loss": 3.66125750541687, + "step": 3318 + }, + { + "epoch": 46.09781659388646, + "grad_norm": 0.005628711078315973, + "learning_rate": 7.73192969227491e-05, + "loss": 3.65299129486084, + "step": 3319 + }, + { + "epoch": 46.1117903930131, + "grad_norm": 0.005935952067375183, + "learning_rate": 7.698607203244189e-05, + "loss": 3.63519287109375, + "step": 3320 + }, + { + "epoch": 46.12576419213974, + "grad_norm": 0.005684727802872658, + "learning_rate": 7.665400572398285e-05, + "loss": 3.6002516746520996, + "step": 3321 + }, + { + "epoch": 46.13973799126637, + "grad_norm": 0.005795048549771309, + "learning_rate": 7.632309826491177e-05, + "loss": 3.6974287033081055, + "step": 3322 + }, + { + "epoch": 46.15371179039301, + "grad_norm": 0.0059766992926597595, + "learning_rate": 7.599334992183438e-05, + "loss": 3.672163486480713, + "step": 3323 + }, + { + "epoch": 46.16768558951965, + "grad_norm": 0.005643186625093222, + "learning_rate": 7.566476096042397e-05, + "loss": 3.6390175819396973, + "step": 3324 + }, + { + "epoch": 46.18165938864629, + "grad_norm": 0.0058948020450770855, + "learning_rate": 7.533733164541767e-05, + "loss": 3.635690689086914, + "step": 3325 + }, + { + "epoch": 46.19563318777293, + "grad_norm": 0.005942242220044136, + "learning_rate": 7.501106224062e-05, + "loss": 3.6397767066955566, + "step": 3326 + }, + { + "epoch": 46.209606986899566, + "grad_norm": 0.0054268259555101395, + "learning_rate": 7.468595300890026e-05, + "loss": 3.634727954864502, + "step": 3327 + }, + { + "epoch": 46.223580786026204, + "grad_norm": 0.006279955618083477, + "learning_rate": 7.43620042121929e-05, + "loss": 3.6551690101623535, + "step": 3328 + }, + { + "epoch": 46.237554585152836, + "grad_norm": 0.005737216677516699, + "learning_rate": 7.403921611149778e-05, + "loss": 3.665435314178467, + "step": 3329 + }, + { + "epoch": 46.251528384279474, + "grad_norm": 0.005809346679598093, + "learning_rate": 7.371758896687908e-05, + "loss": 3.645390033721924, + "step": 3330 + }, + { + "epoch": 46.26550218340611, + "grad_norm": 0.005774138495326042, + "learning_rate": 7.339712303746658e-05, + "loss": 3.6289312839508057, + "step": 3331 + }, + { + "epoch": 46.27947598253275, + "grad_norm": 0.005683782510459423, + "learning_rate": 7.30778185814535e-05, + "loss": 3.6434290409088135, + "step": 3332 + }, + { + "epoch": 46.29344978165939, + "grad_norm": 0.006072668358683586, + "learning_rate": 7.275967585609755e-05, + "loss": 3.6633148193359375, + "step": 3333 + }, + { + "epoch": 46.30742358078603, + "grad_norm": 0.005564856808632612, + "learning_rate": 7.244269511772106e-05, + "loss": 3.617435932159424, + "step": 3334 + }, + { + "epoch": 46.32139737991266, + "grad_norm": 0.005909780040383339, + "learning_rate": 7.212687662170915e-05, + "loss": 3.6613821983337402, + "step": 3335 + }, + { + "epoch": 46.3353711790393, + "grad_norm": 0.005920986644923687, + "learning_rate": 7.181222062251155e-05, + "loss": 3.6254312992095947, + "step": 3336 + }, + { + "epoch": 46.34934497816594, + "grad_norm": 0.005616472102701664, + "learning_rate": 7.14987273736407e-05, + "loss": 3.653961658477783, + "step": 3337 + }, + { + "epoch": 46.36331877729258, + "grad_norm": 0.006085917819291353, + "learning_rate": 7.118639712767222e-05, + "loss": 3.6716418266296387, + "step": 3338 + }, + { + "epoch": 46.377292576419215, + "grad_norm": 0.0057594352401793, + "learning_rate": 7.087523013624534e-05, + "loss": 3.66914701461792, + "step": 3339 + }, + { + "epoch": 46.391266375545854, + "grad_norm": 0.005555839743465185, + "learning_rate": 7.056522665006162e-05, + "loss": 3.674562692642212, + "step": 3340 + }, + { + "epoch": 46.40524017467249, + "grad_norm": 0.005709810648113489, + "learning_rate": 7.025638691888476e-05, + "loss": 3.664320468902588, + "step": 3341 + }, + { + "epoch": 46.419213973799124, + "grad_norm": 0.005785972345620394, + "learning_rate": 6.994871119154197e-05, + "loss": 3.6416711807250977, + "step": 3342 + }, + { + "epoch": 46.43318777292576, + "grad_norm": 0.0058121830224990845, + "learning_rate": 6.964219971592198e-05, + "loss": 3.654954433441162, + "step": 3343 + }, + { + "epoch": 46.4471615720524, + "grad_norm": 0.005938741844147444, + "learning_rate": 6.933685273897533e-05, + "loss": 3.6046786308288574, + "step": 3344 + }, + { + "epoch": 46.46113537117904, + "grad_norm": 0.005746894516050816, + "learning_rate": 6.903267050671447e-05, + "loss": 3.6525707244873047, + "step": 3345 + }, + { + "epoch": 46.47510917030568, + "grad_norm": 0.005644070915877819, + "learning_rate": 6.872965326421365e-05, + "loss": 3.652769088745117, + "step": 3346 + }, + { + "epoch": 46.48908296943232, + "grad_norm": 0.006110438145697117, + "learning_rate": 6.84278012556088e-05, + "loss": 3.669772148132324, + "step": 3347 + }, + { + "epoch": 46.50305676855895, + "grad_norm": 0.0055745323188602924, + "learning_rate": 6.81271147240959e-05, + "loss": 3.6494526863098145, + "step": 3348 + }, + { + "epoch": 46.51703056768559, + "grad_norm": 0.005900663323700428, + "learning_rate": 6.78275939119328e-05, + "loss": 3.6548643112182617, + "step": 3349 + }, + { + "epoch": 46.531004366812226, + "grad_norm": 0.00568796694278717, + "learning_rate": 6.752923906043862e-05, + "loss": 3.6537368297576904, + "step": 3350 + }, + { + "epoch": 46.544978165938865, + "grad_norm": 0.005612431094050407, + "learning_rate": 6.723205040999193e-05, + "loss": 3.6221089363098145, + "step": 3351 + }, + { + "epoch": 46.5589519650655, + "grad_norm": 0.005935634020715952, + "learning_rate": 6.693602820003248e-05, + "loss": 3.6739964485168457, + "step": 3352 + }, + { + "epoch": 46.57292576419214, + "grad_norm": 0.005767039954662323, + "learning_rate": 6.66411726690598e-05, + "loss": 3.6298999786376953, + "step": 3353 + }, + { + "epoch": 46.58689956331878, + "grad_norm": 0.005834923591464758, + "learning_rate": 6.634748405463404e-05, + "loss": 3.6653504371643066, + "step": 3354 + }, + { + "epoch": 46.60087336244541, + "grad_norm": 0.005940297152847052, + "learning_rate": 6.605496259337425e-05, + "loss": 3.6638693809509277, + "step": 3355 + }, + { + "epoch": 46.61484716157205, + "grad_norm": 0.00571000250056386, + "learning_rate": 6.576360852096006e-05, + "loss": 3.639021873474121, + "step": 3356 + }, + { + "epoch": 46.62882096069869, + "grad_norm": 0.0057969787158071995, + "learning_rate": 6.547342207213038e-05, + "loss": 3.6597180366516113, + "step": 3357 + }, + { + "epoch": 46.64279475982533, + "grad_norm": 0.005795357748866081, + "learning_rate": 6.518440348068267e-05, + "loss": 3.6444644927978516, + "step": 3358 + }, + { + "epoch": 46.65676855895197, + "grad_norm": 0.005731288809329271, + "learning_rate": 6.489655297947441e-05, + "loss": 3.668120861053467, + "step": 3359 + }, + { + "epoch": 46.670742358078606, + "grad_norm": 0.0058074090629816055, + "learning_rate": 6.460987080042126e-05, + "loss": 3.639407157897949, + "step": 3360 + }, + { + "epoch": 46.68471615720524, + "grad_norm": 0.005735141225159168, + "learning_rate": 6.432435717449837e-05, + "loss": 3.6410751342773438, + "step": 3361 + }, + { + "epoch": 46.698689956331876, + "grad_norm": 0.005885255988687277, + "learning_rate": 6.404001233173828e-05, + "loss": 3.658564567565918, + "step": 3362 + }, + { + "epoch": 46.712663755458514, + "grad_norm": 0.005977948661893606, + "learning_rate": 6.375683650123268e-05, + "loss": 3.688422918319702, + "step": 3363 + }, + { + "epoch": 46.72663755458515, + "grad_norm": 0.005629570223391056, + "learning_rate": 6.347482991113151e-05, + "loss": 3.6477694511413574, + "step": 3364 + }, + { + "epoch": 46.74061135371179, + "grad_norm": 0.005695016123354435, + "learning_rate": 6.319399278864202e-05, + "loss": 3.6869497299194336, + "step": 3365 + }, + { + "epoch": 46.75458515283843, + "grad_norm": 0.0057486798614263535, + "learning_rate": 6.291432536003e-05, + "loss": 3.6609716415405273, + "step": 3366 + }, + { + "epoch": 46.76855895196506, + "grad_norm": 0.005637758877128363, + "learning_rate": 6.263582785061794e-05, + "loss": 3.644331932067871, + "step": 3367 + }, + { + "epoch": 46.7825327510917, + "grad_norm": 0.005615375470370054, + "learning_rate": 6.235850048478689e-05, + "loss": 3.6516261100769043, + "step": 3368 + }, + { + "epoch": 46.79650655021834, + "grad_norm": 0.005714157596230507, + "learning_rate": 6.208234348597384e-05, + "loss": 3.6313061714172363, + "step": 3369 + }, + { + "epoch": 46.81048034934498, + "grad_norm": 0.005742870736867189, + "learning_rate": 6.180735707667399e-05, + "loss": 3.658287525177002, + "step": 3370 + }, + { + "epoch": 46.82445414847162, + "grad_norm": 0.005571036599576473, + "learning_rate": 6.153354147843854e-05, + "loss": 3.6537227630615234, + "step": 3371 + }, + { + "epoch": 46.838427947598255, + "grad_norm": 0.005676159635186195, + "learning_rate": 6.126089691187606e-05, + "loss": 3.6534817218780518, + "step": 3372 + }, + { + "epoch": 46.852401746724894, + "grad_norm": 0.00567635940387845, + "learning_rate": 6.098942359665141e-05, + "loss": 3.62833309173584, + "step": 3373 + }, + { + "epoch": 46.866375545851525, + "grad_norm": 0.005548842716962099, + "learning_rate": 6.071912175148553e-05, + "loss": 3.6426186561584473, + "step": 3374 + }, + { + "epoch": 46.880349344978164, + "grad_norm": 0.0056854793801903725, + "learning_rate": 6.0449991594155694e-05, + "loss": 3.670558452606201, + "step": 3375 + }, + { + "epoch": 46.8943231441048, + "grad_norm": 0.005691561382263899, + "learning_rate": 6.0182033341495317e-05, + "loss": 3.6603145599365234, + "step": 3376 + }, + { + "epoch": 46.90829694323144, + "grad_norm": 0.005610436201095581, + "learning_rate": 5.991524720939341e-05, + "loss": 3.6651010513305664, + "step": 3377 + }, + { + "epoch": 46.92227074235808, + "grad_norm": 0.005553451832383871, + "learning_rate": 5.9649633412794984e-05, + "loss": 3.680150032043457, + "step": 3378 + }, + { + "epoch": 46.93624454148472, + "grad_norm": 0.005749273579567671, + "learning_rate": 5.93851921657e-05, + "loss": 3.6609292030334473, + "step": 3379 + }, + { + "epoch": 46.95021834061135, + "grad_norm": 0.005574697162955999, + "learning_rate": 5.91219236811644e-05, + "loss": 3.7022757530212402, + "step": 3380 + }, + { + "epoch": 46.96419213973799, + "grad_norm": 0.005650301463901997, + "learning_rate": 5.885982817129827e-05, + "loss": 3.71413516998291, + "step": 3381 + }, + { + "epoch": 46.97816593886463, + "grad_norm": 0.005744843743741512, + "learning_rate": 5.859890584726769e-05, + "loss": 3.6550443172454834, + "step": 3382 + }, + { + "epoch": 46.992139737991266, + "grad_norm": 0.00542390625923872, + "learning_rate": 5.833915691929282e-05, + "loss": 3.6864724159240723, + "step": 3383 + }, + { + "epoch": 47.0, + "grad_norm": 0.007102064788341522, + "learning_rate": 5.808058159664905e-05, + "loss": 3.681469440460205, + "step": 3384 + }, + { + "epoch": 47.0, + "eval_loss": 4.382168769836426, + "eval_runtime": 52.2293, + "eval_samples_per_second": 46.755, + "eval_steps_per_second": 1.474, + "step": 3384 + }, + { + "epoch": 47.01397379912664, + "grad_norm": 0.005845201667398214, + "learning_rate": 5.782318008766519e-05, + "loss": 3.63492751121521, + "step": 3385 + }, + { + "epoch": 47.02794759825328, + "grad_norm": 0.005614436697214842, + "learning_rate": 5.75669525997254e-05, + "loss": 3.629580020904541, + "step": 3386 + }, + { + "epoch": 47.041921397379916, + "grad_norm": 0.0055298879742622375, + "learning_rate": 5.731189933926781e-05, + "loss": 3.6301462650299072, + "step": 3387 + }, + { + "epoch": 47.05589519650655, + "grad_norm": 0.005644683260470629, + "learning_rate": 5.7058020511783706e-05, + "loss": 3.6150715351104736, + "step": 3388 + }, + { + "epoch": 47.069868995633186, + "grad_norm": 0.005629776511341333, + "learning_rate": 5.6805316321818886e-05, + "loss": 3.6537346839904785, + "step": 3389 + }, + { + "epoch": 47.083842794759825, + "grad_norm": 0.005810001865029335, + "learning_rate": 5.655378697297258e-05, + "loss": 3.64941143989563, + "step": 3390 + }, + { + "epoch": 47.09781659388646, + "grad_norm": 0.0057138605043292046, + "learning_rate": 5.630343266789741e-05, + "loss": 3.6648316383361816, + "step": 3391 + }, + { + "epoch": 47.1117903930131, + "grad_norm": 0.005495130084455013, + "learning_rate": 5.605425360829931e-05, + "loss": 3.619474411010742, + "step": 3392 + }, + { + "epoch": 47.12576419213974, + "grad_norm": 0.005678716115653515, + "learning_rate": 5.580624999493715e-05, + "loss": 3.6268370151519775, + "step": 3393 + }, + { + "epoch": 47.13973799126637, + "grad_norm": 0.0057465555146336555, + "learning_rate": 5.5559422027623556e-05, + "loss": 3.6618542671203613, + "step": 3394 + }, + { + "epoch": 47.15371179039301, + "grad_norm": 0.005927131045609713, + "learning_rate": 5.5313769905222756e-05, + "loss": 3.6629247665405273, + "step": 3395 + }, + { + "epoch": 47.16768558951965, + "grad_norm": 0.0056359837763011456, + "learning_rate": 5.506929382565235e-05, + "loss": 3.638967990875244, + "step": 3396 + }, + { + "epoch": 47.18165938864629, + "grad_norm": 0.005787394475191832, + "learning_rate": 5.482599398588257e-05, + "loss": 3.6664178371429443, + "step": 3397 + }, + { + "epoch": 47.19563318777293, + "grad_norm": 0.0054801893420517445, + "learning_rate": 5.458387058193552e-05, + "loss": 3.6431665420532227, + "step": 3398 + }, + { + "epoch": 47.209606986899566, + "grad_norm": 0.005605428479611874, + "learning_rate": 5.434292380888563e-05, + "loss": 3.6501848697662354, + "step": 3399 + }, + { + "epoch": 47.223580786026204, + "grad_norm": 0.005695626139640808, + "learning_rate": 5.410315386085934e-05, + "loss": 3.6499099731445312, + "step": 3400 + }, + { + "epoch": 47.237554585152836, + "grad_norm": 0.005612112116068602, + "learning_rate": 5.386456093103495e-05, + "loss": 3.6890406608581543, + "step": 3401 + }, + { + "epoch": 47.251528384279474, + "grad_norm": 0.005486638285219669, + "learning_rate": 5.362714521164272e-05, + "loss": 3.6657395362854004, + "step": 3402 + }, + { + "epoch": 47.26550218340611, + "grad_norm": 0.005863480269908905, + "learning_rate": 5.3390906893963974e-05, + "loss": 3.666640281677246, + "step": 3403 + }, + { + "epoch": 47.27947598253275, + "grad_norm": 0.00578767666593194, + "learning_rate": 5.315584616833154e-05, + "loss": 3.6408395767211914, + "step": 3404 + }, + { + "epoch": 47.29344978165939, + "grad_norm": 0.005513481795787811, + "learning_rate": 5.2921963224130086e-05, + "loss": 3.661557197570801, + "step": 3405 + }, + { + "epoch": 47.30742358078603, + "grad_norm": 0.0056565022096037865, + "learning_rate": 5.268925824979431e-05, + "loss": 3.646548271179199, + "step": 3406 + }, + { + "epoch": 47.32139737991266, + "grad_norm": 0.0057138074189424515, + "learning_rate": 5.2457731432810745e-05, + "loss": 3.654798984527588, + "step": 3407 + }, + { + "epoch": 47.3353711790393, + "grad_norm": 0.005857877433300018, + "learning_rate": 5.222738295971629e-05, + "loss": 3.6370182037353516, + "step": 3408 + }, + { + "epoch": 47.34934497816594, + "grad_norm": 0.005622264463454485, + "learning_rate": 5.199821301609856e-05, + "loss": 3.6596436500549316, + "step": 3409 + }, + { + "epoch": 47.36331877729258, + "grad_norm": 0.005421077832579613, + "learning_rate": 5.1770221786595654e-05, + "loss": 3.638583183288574, + "step": 3410 + }, + { + "epoch": 47.377292576419215, + "grad_norm": 0.0054649352096021175, + "learning_rate": 5.154340945489587e-05, + "loss": 3.6493330001831055, + "step": 3411 + }, + { + "epoch": 47.391266375545854, + "grad_norm": 0.005776858888566494, + "learning_rate": 5.1317776203737905e-05, + "loss": 3.6507070064544678, + "step": 3412 + }, + { + "epoch": 47.40524017467249, + "grad_norm": 0.005619521252810955, + "learning_rate": 5.109332221491043e-05, + "loss": 3.635849952697754, + "step": 3413 + }, + { + "epoch": 47.419213973799124, + "grad_norm": 0.0055708554573357105, + "learning_rate": 5.0870047669251846e-05, + "loss": 3.68941068649292, + "step": 3414 + }, + { + "epoch": 47.43318777292576, + "grad_norm": 0.005647908430546522, + "learning_rate": 5.06479527466506e-05, + "loss": 3.666015625, + "step": 3415 + }, + { + "epoch": 47.4471615720524, + "grad_norm": 0.005934152286499739, + "learning_rate": 5.0427037626044326e-05, + "loss": 3.6556167602539062, + "step": 3416 + }, + { + "epoch": 47.46113537117904, + "grad_norm": 0.005791633389890194, + "learning_rate": 5.020730248542068e-05, + "loss": 3.6541948318481445, + "step": 3417 + }, + { + "epoch": 47.47510917030568, + "grad_norm": 0.00573629979044199, + "learning_rate": 4.998874750181586e-05, + "loss": 3.626328468322754, + "step": 3418 + }, + { + "epoch": 47.48908296943232, + "grad_norm": 0.005624848417937756, + "learning_rate": 4.9771372851315965e-05, + "loss": 3.6370186805725098, + "step": 3419 + }, + { + "epoch": 47.50305676855895, + "grad_norm": 0.005743588786572218, + "learning_rate": 4.955517870905577e-05, + "loss": 3.647320032119751, + "step": 3420 + }, + { + "epoch": 47.51703056768559, + "grad_norm": 0.005720328539609909, + "learning_rate": 4.934016524921894e-05, + "loss": 3.607609272003174, + "step": 3421 + }, + { + "epoch": 47.531004366812226, + "grad_norm": 0.005619847681373358, + "learning_rate": 4.912633264503801e-05, + "loss": 3.676149845123291, + "step": 3422 + }, + { + "epoch": 47.544978165938865, + "grad_norm": 0.00549230445176363, + "learning_rate": 4.8913681068793916e-05, + "loss": 3.6969799995422363, + "step": 3423 + }, + { + "epoch": 47.5589519650655, + "grad_norm": 0.0055993059650063515, + "learning_rate": 4.8702210691816596e-05, + "loss": 3.6594631671905518, + "step": 3424 + }, + { + "epoch": 47.57292576419214, + "grad_norm": 0.005832983180880547, + "learning_rate": 4.84919216844834e-05, + "loss": 3.6583518981933594, + "step": 3425 + }, + { + "epoch": 47.58689956331878, + "grad_norm": 0.005687489174306393, + "learning_rate": 4.828281421622057e-05, + "loss": 3.6750564575195312, + "step": 3426 + }, + { + "epoch": 47.60087336244541, + "grad_norm": 0.0055601997300982475, + "learning_rate": 4.807488845550223e-05, + "loss": 3.6731209754943848, + "step": 3427 + }, + { + "epoch": 47.61484716157205, + "grad_norm": 0.005822640378028154, + "learning_rate": 4.786814456985036e-05, + "loss": 3.6643764972686768, + "step": 3428 + }, + { + "epoch": 47.62882096069869, + "grad_norm": 0.0057311588898301125, + "learning_rate": 4.7662582725835025e-05, + "loss": 3.6473302841186523, + "step": 3429 + }, + { + "epoch": 47.64279475982533, + "grad_norm": 0.005722011439502239, + "learning_rate": 4.745820308907318e-05, + "loss": 3.6761927604675293, + "step": 3430 + }, + { + "epoch": 47.65676855895197, + "grad_norm": 0.005630903877317905, + "learning_rate": 4.725500582423046e-05, + "loss": 3.668677806854248, + "step": 3431 + }, + { + "epoch": 47.670742358078606, + "grad_norm": 0.005585893988609314, + "learning_rate": 4.7052991095018625e-05, + "loss": 3.6321611404418945, + "step": 3432 + }, + { + "epoch": 47.68471615720524, + "grad_norm": 0.005564410239458084, + "learning_rate": 4.6852159064197785e-05, + "loss": 3.649759531021118, + "step": 3433 + }, + { + "epoch": 47.698689956331876, + "grad_norm": 0.005554834380745888, + "learning_rate": 4.665250989357429e-05, + "loss": 3.6571452617645264, + "step": 3434 + }, + { + "epoch": 47.712663755458514, + "grad_norm": 0.005567575339227915, + "learning_rate": 4.645404374400205e-05, + "loss": 3.653958320617676, + "step": 3435 + }, + { + "epoch": 47.72663755458515, + "grad_norm": 0.00546268792822957, + "learning_rate": 4.62567607753817e-05, + "loss": 3.6481666564941406, + "step": 3436 + }, + { + "epoch": 47.74061135371179, + "grad_norm": 0.0055222418159246445, + "learning_rate": 4.606066114666047e-05, + "loss": 3.6125011444091797, + "step": 3437 + }, + { + "epoch": 47.75458515283843, + "grad_norm": 0.0054062786512076855, + "learning_rate": 4.586574501583229e-05, + "loss": 3.6090235710144043, + "step": 3438 + }, + { + "epoch": 47.76855895196506, + "grad_norm": 0.005483254324644804, + "learning_rate": 4.5672012539937675e-05, + "loss": 3.647038459777832, + "step": 3439 + }, + { + "epoch": 47.7825327510917, + "grad_norm": 0.005621953401714563, + "learning_rate": 4.547946387506337e-05, + "loss": 3.64411997795105, + "step": 3440 + }, + { + "epoch": 47.79650655021834, + "grad_norm": 0.005623196717351675, + "learning_rate": 4.5288099176342185e-05, + "loss": 3.618316173553467, + "step": 3441 + }, + { + "epoch": 47.81048034934498, + "grad_norm": 0.005601753015071154, + "learning_rate": 4.509791859795321e-05, + "loss": 3.644348621368408, + "step": 3442 + }, + { + "epoch": 47.82445414847162, + "grad_norm": 0.005769635550677776, + "learning_rate": 4.490892229312188e-05, + "loss": 3.6343932151794434, + "step": 3443 + }, + { + "epoch": 47.838427947598255, + "grad_norm": 0.005564125254750252, + "learning_rate": 4.472111041411859e-05, + "loss": 3.6716389656066895, + "step": 3444 + }, + { + "epoch": 47.852401746724894, + "grad_norm": 0.0054967706091701984, + "learning_rate": 4.453448311226028e-05, + "loss": 3.6779255867004395, + "step": 3445 + }, + { + "epoch": 47.866375545851525, + "grad_norm": 0.005589650943875313, + "learning_rate": 4.4349040537909155e-05, + "loss": 3.63474702835083, + "step": 3446 + }, + { + "epoch": 47.880349344978164, + "grad_norm": 0.005526749417185783, + "learning_rate": 4.416478284047299e-05, + "loss": 3.65665864944458, + "step": 3447 + }, + { + "epoch": 47.8943231441048, + "grad_norm": 0.005574159789830446, + "learning_rate": 4.398171016840464e-05, + "loss": 3.6104025840759277, + "step": 3448 + }, + { + "epoch": 47.90829694323144, + "grad_norm": 0.00546378456056118, + "learning_rate": 4.379982266920288e-05, + "loss": 3.6659083366394043, + "step": 3449 + }, + { + "epoch": 47.92227074235808, + "grad_norm": 0.005439163185656071, + "learning_rate": 4.361912048941073e-05, + "loss": 3.6359872817993164, + "step": 3450 + }, + { + "epoch": 47.93624454148472, + "grad_norm": 0.005478638224303722, + "learning_rate": 4.343960377461697e-05, + "loss": 3.63889217376709, + "step": 3451 + }, + { + "epoch": 47.95021834061135, + "grad_norm": 0.005485034082084894, + "learning_rate": 4.326127266945479e-05, + "loss": 3.6665172576904297, + "step": 3452 + }, + { + "epoch": 47.96419213973799, + "grad_norm": 0.006088086403906345, + "learning_rate": 4.3084127317602475e-05, + "loss": 3.684257745742798, + "step": 3453 + }, + { + "epoch": 47.97816593886463, + "grad_norm": 0.005437501240521669, + "learning_rate": 4.290816786178288e-05, + "loss": 3.6222567558288574, + "step": 3454 + }, + { + "epoch": 47.992139737991266, + "grad_norm": 0.005640770774334669, + "learning_rate": 4.2733394443763146e-05, + "loss": 3.651585578918457, + "step": 3455 + }, + { + "epoch": 48.0, + "grad_norm": 0.006981590297073126, + "learning_rate": 4.255980720435515e-05, + "loss": 3.6314165592193604, + "step": 3456 + }, + { + "epoch": 48.0, + "eval_loss": 4.384180068969727, + "eval_runtime": 51.5098, + "eval_samples_per_second": 47.408, + "eval_steps_per_second": 1.495, + "step": 3456 + }, + { + "epoch": 48.01397379912664, + "grad_norm": 0.0055017475970089436, + "learning_rate": 4.238740628341503e-05, + "loss": 3.588742733001709, + "step": 3457 + }, + { + "epoch": 48.02794759825328, + "grad_norm": 0.005516123026609421, + "learning_rate": 4.22161918198432e-05, + "loss": 3.6423656940460205, + "step": 3458 + }, + { + "epoch": 48.041921397379916, + "grad_norm": 0.005525878164917231, + "learning_rate": 4.204616395158382e-05, + "loss": 3.6513490676879883, + "step": 3459 + }, + { + "epoch": 48.05589519650655, + "grad_norm": 0.005577936768531799, + "learning_rate": 4.1877322815625474e-05, + "loss": 3.6643600463867188, + "step": 3460 + }, + { + "epoch": 48.069868995633186, + "grad_norm": 0.005561518482863903, + "learning_rate": 4.170966854800052e-05, + "loss": 3.66424298286438, + "step": 3461 + }, + { + "epoch": 48.083842794759825, + "grad_norm": 0.005620717070996761, + "learning_rate": 4.154320128378459e-05, + "loss": 3.6455154418945312, + "step": 3462 + }, + { + "epoch": 48.09781659388646, + "grad_norm": 0.00556279718875885, + "learning_rate": 4.137792115709756e-05, + "loss": 3.634735584259033, + "step": 3463 + }, + { + "epoch": 48.1117903930131, + "grad_norm": 0.005605190526694059, + "learning_rate": 4.1213828301102755e-05, + "loss": 3.610032081604004, + "step": 3464 + }, + { + "epoch": 48.12576419213974, + "grad_norm": 0.005742431152611971, + "learning_rate": 4.1050922848006605e-05, + "loss": 3.6297264099121094, + "step": 3465 + }, + { + "epoch": 48.13973799126637, + "grad_norm": 0.005596380680799484, + "learning_rate": 4.088920492905929e-05, + "loss": 3.678337574005127, + "step": 3466 + }, + { + "epoch": 48.15371179039301, + "grad_norm": 0.0056046731770038605, + "learning_rate": 4.0728674674553794e-05, + "loss": 3.6512162685394287, + "step": 3467 + }, + { + "epoch": 48.16768558951965, + "grad_norm": 0.005494440905749798, + "learning_rate": 4.0569332213826826e-05, + "loss": 3.6533617973327637, + "step": 3468 + }, + { + "epoch": 48.18165938864629, + "grad_norm": 0.005756898783147335, + "learning_rate": 4.041117767525725e-05, + "loss": 3.649506092071533, + "step": 3469 + }, + { + "epoch": 48.19563318777293, + "grad_norm": 0.005806957837194204, + "learning_rate": 4.025421118626767e-05, + "loss": 3.695404529571533, + "step": 3470 + }, + { + "epoch": 48.209606986899566, + "grad_norm": 0.005643757060170174, + "learning_rate": 4.009843287332282e-05, + "loss": 3.656587600708008, + "step": 3471 + }, + { + "epoch": 48.223580786026204, + "grad_norm": 0.005467074923217297, + "learning_rate": 3.9943842861930864e-05, + "loss": 3.671684741973877, + "step": 3472 + }, + { + "epoch": 48.237554585152836, + "grad_norm": 0.005573616363108158, + "learning_rate": 3.9790441276642075e-05, + "loss": 3.661931037902832, + "step": 3473 + }, + { + "epoch": 48.251528384279474, + "grad_norm": 0.005787190515547991, + "learning_rate": 3.963822824104919e-05, + "loss": 3.629055976867676, + "step": 3474 + }, + { + "epoch": 48.26550218340611, + "grad_norm": 0.005627274047583342, + "learning_rate": 3.948720387778753e-05, + "loss": 3.6431307792663574, + "step": 3475 + }, + { + "epoch": 48.27947598253275, + "grad_norm": 0.005619972012937069, + "learning_rate": 3.933736830853473e-05, + "loss": 3.674844741821289, + "step": 3476 + }, + { + "epoch": 48.29344978165939, + "grad_norm": 0.0055321259424090385, + "learning_rate": 3.918872165401068e-05, + "loss": 3.6465535163879395, + "step": 3477 + }, + { + "epoch": 48.30742358078603, + "grad_norm": 0.005573419854044914, + "learning_rate": 3.904126403397724e-05, + "loss": 3.616992950439453, + "step": 3478 + }, + { + "epoch": 48.32139737991266, + "grad_norm": 0.005475590471178293, + "learning_rate": 3.889499556723837e-05, + "loss": 3.624130964279175, + "step": 3479 + }, + { + "epoch": 48.3353711790393, + "grad_norm": 0.005683086346834898, + "learning_rate": 3.874991637163997e-05, + "loss": 3.632981538772583, + "step": 3480 + }, + { + "epoch": 48.34934497816594, + "grad_norm": 0.005622202064841986, + "learning_rate": 3.860602656406961e-05, + "loss": 3.6468749046325684, + "step": 3481 + }, + { + "epoch": 48.36331877729258, + "grad_norm": 0.005536191631108522, + "learning_rate": 3.84633262604566e-05, + "loss": 3.633355140686035, + "step": 3482 + }, + { + "epoch": 48.377292576419215, + "grad_norm": 0.00547434389591217, + "learning_rate": 3.8321815575772385e-05, + "loss": 3.6079788208007812, + "step": 3483 + }, + { + "epoch": 48.391266375545854, + "grad_norm": 0.005573789123445749, + "learning_rate": 3.818149462402937e-05, + "loss": 3.6123361587524414, + "step": 3484 + }, + { + "epoch": 48.40524017467249, + "grad_norm": 0.00561481062322855, + "learning_rate": 3.8042363518281414e-05, + "loss": 3.664466142654419, + "step": 3485 + }, + { + "epoch": 48.419213973799124, + "grad_norm": 0.005487846210598946, + "learning_rate": 3.790442237062417e-05, + "loss": 3.655763626098633, + "step": 3486 + }, + { + "epoch": 48.43318777292576, + "grad_norm": 0.005416626576334238, + "learning_rate": 3.7767671292194405e-05, + "loss": 3.634082317352295, + "step": 3487 + }, + { + "epoch": 48.4471615720524, + "grad_norm": 0.005675632040947676, + "learning_rate": 3.76321103931697e-05, + "loss": 3.6449050903320312, + "step": 3488 + }, + { + "epoch": 48.46113537117904, + "grad_norm": 0.0055493442341685295, + "learning_rate": 3.749773978276924e-05, + "loss": 3.6530330181121826, + "step": 3489 + }, + { + "epoch": 48.47510917030568, + "grad_norm": 0.005522808991372585, + "learning_rate": 3.7364559569253025e-05, + "loss": 3.6732282638549805, + "step": 3490 + }, + { + "epoch": 48.48908296943232, + "grad_norm": 0.005541914142668247, + "learning_rate": 3.7232569859921836e-05, + "loss": 3.6114039421081543, + "step": 3491 + }, + { + "epoch": 48.50305676855895, + "grad_norm": 0.005627194419503212, + "learning_rate": 3.710177076111743e-05, + "loss": 3.6930434703826904, + "step": 3492 + }, + { + "epoch": 48.51703056768559, + "grad_norm": 0.005525750108063221, + "learning_rate": 3.697216237822202e-05, + "loss": 3.6784677505493164, + "step": 3493 + }, + { + "epoch": 48.531004366812226, + "grad_norm": 0.005604377947747707, + "learning_rate": 3.6843744815659275e-05, + "loss": 3.631898880004883, + "step": 3494 + }, + { + "epoch": 48.544978165938865, + "grad_norm": 0.005518812220543623, + "learning_rate": 3.671651817689269e-05, + "loss": 3.663426399230957, + "step": 3495 + }, + { + "epoch": 48.5589519650655, + "grad_norm": 0.005526282824575901, + "learning_rate": 3.6590482564426204e-05, + "loss": 3.630021572113037, + "step": 3496 + }, + { + "epoch": 48.57292576419214, + "grad_norm": 0.005505071487277746, + "learning_rate": 3.646563807980474e-05, + "loss": 3.6589057445526123, + "step": 3497 + }, + { + "epoch": 48.58689956331878, + "grad_norm": 0.00532246520742774, + "learning_rate": 3.6341984823613016e-05, + "loss": 3.640037775039673, + "step": 3498 + }, + { + "epoch": 48.60087336244541, + "grad_norm": 0.005556885153055191, + "learning_rate": 3.6219522895476545e-05, + "loss": 3.6187455654144287, + "step": 3499 + }, + { + "epoch": 48.61484716157205, + "grad_norm": 0.005715913604944944, + "learning_rate": 3.609825239406033e-05, + "loss": 3.633690118789673, + "step": 3500 + }, + { + "epoch": 48.62882096069869, + "grad_norm": 0.00542859872803092, + "learning_rate": 3.5978173417069984e-05, + "loss": 3.6094117164611816, + "step": 3501 + }, + { + "epoch": 48.64279475982533, + "grad_norm": 0.005408013239502907, + "learning_rate": 3.585928606125113e-05, + "loss": 3.680063247680664, + "step": 3502 + }, + { + "epoch": 48.65676855895197, + "grad_norm": 0.005575202405452728, + "learning_rate": 3.574159042238883e-05, + "loss": 3.6290431022644043, + "step": 3503 + }, + { + "epoch": 48.670742358078606, + "grad_norm": 0.005622019059956074, + "learning_rate": 3.562508659530864e-05, + "loss": 3.6656417846679688, + "step": 3504 + }, + { + "epoch": 48.68471615720524, + "grad_norm": 0.005383938550949097, + "learning_rate": 3.5509774673875424e-05, + "loss": 3.638788938522339, + "step": 3505 + }, + { + "epoch": 48.698689956331876, + "grad_norm": 0.005429980810731649, + "learning_rate": 3.539565475099401e-05, + "loss": 3.660780191421509, + "step": 3506 + }, + { + "epoch": 48.712663755458514, + "grad_norm": 0.005457386840134859, + "learning_rate": 3.528272691860856e-05, + "loss": 3.6496505737304688, + "step": 3507 + }, + { + "epoch": 48.72663755458515, + "grad_norm": 0.005568833090364933, + "learning_rate": 3.517099126770333e-05, + "loss": 3.657479763031006, + "step": 3508 + }, + { + "epoch": 48.74061135371179, + "grad_norm": 0.005413518752902746, + "learning_rate": 3.506044788830128e-05, + "loss": 3.681161642074585, + "step": 3509 + }, + { + "epoch": 48.75458515283843, + "grad_norm": 0.005438497755676508, + "learning_rate": 3.4951096869465646e-05, + "loss": 3.6673734188079834, + "step": 3510 + }, + { + "epoch": 48.76855895196506, + "grad_norm": 0.005298427771776915, + "learning_rate": 3.484293829929833e-05, + "loss": 3.6321797370910645, + "step": 3511 + }, + { + "epoch": 48.7825327510917, + "grad_norm": 0.0056403265334665775, + "learning_rate": 3.473597226494069e-05, + "loss": 3.625696897506714, + "step": 3512 + }, + { + "epoch": 48.79650655021834, + "grad_norm": 0.005573336500674486, + "learning_rate": 3.463019885257344e-05, + "loss": 3.656543254852295, + "step": 3513 + }, + { + "epoch": 48.81048034934498, + "grad_norm": 0.005373518913984299, + "learning_rate": 3.4525618147416384e-05, + "loss": 3.6582937240600586, + "step": 3514 + }, + { + "epoch": 48.82445414847162, + "grad_norm": 0.005456839222460985, + "learning_rate": 3.442223023372804e-05, + "loss": 3.6304612159729004, + "step": 3515 + }, + { + "epoch": 48.838427947598255, + "grad_norm": 0.0055037084966897964, + "learning_rate": 3.432003519480653e-05, + "loss": 3.6726675033569336, + "step": 3516 + }, + { + "epoch": 48.852401746724894, + "grad_norm": 0.005476827267557383, + "learning_rate": 3.421903311298831e-05, + "loss": 3.6638195514678955, + "step": 3517 + }, + { + "epoch": 48.866375545851525, + "grad_norm": 0.005663580726832151, + "learning_rate": 3.411922406964899e-05, + "loss": 3.6422252655029297, + "step": 3518 + }, + { + "epoch": 48.880349344978164, + "grad_norm": 0.005374568048864603, + "learning_rate": 3.402060814520299e-05, + "loss": 3.6490671634674072, + "step": 3519 + }, + { + "epoch": 48.8943231441048, + "grad_norm": 0.005449265707284212, + "learning_rate": 3.392318541910323e-05, + "loss": 3.6136131286621094, + "step": 3520 + }, + { + "epoch": 48.90829694323144, + "grad_norm": 0.005630240309983492, + "learning_rate": 3.382695596984144e-05, + "loss": 3.677155017852783, + "step": 3521 + }, + { + "epoch": 48.92227074235808, + "grad_norm": 0.005428317468613386, + "learning_rate": 3.373191987494817e-05, + "loss": 3.648580312728882, + "step": 3522 + }, + { + "epoch": 48.93624454148472, + "grad_norm": 0.005498172715306282, + "learning_rate": 3.363807721099197e-05, + "loss": 3.6392035484313965, + "step": 3523 + }, + { + "epoch": 48.95021834061135, + "grad_norm": 0.005650228820741177, + "learning_rate": 3.354542805358054e-05, + "loss": 3.623358964920044, + "step": 3524 + }, + { + "epoch": 48.96419213973799, + "grad_norm": 0.005477136466652155, + "learning_rate": 3.3453972477359395e-05, + "loss": 3.6367993354797363, + "step": 3525 + }, + { + "epoch": 48.97816593886463, + "grad_norm": 0.005379918497055769, + "learning_rate": 3.3363710556012716e-05, + "loss": 3.6611275672912598, + "step": 3526 + }, + { + "epoch": 48.992139737991266, + "grad_norm": 0.00544297369197011, + "learning_rate": 3.327464236226285e-05, + "loss": 3.6238505840301514, + "step": 3527 + }, + { + "epoch": 49.0, + "grad_norm": 0.00707196444272995, + "learning_rate": 3.3186767967870446e-05, + "loss": 3.664491653442383, + "step": 3528 + }, + { + "epoch": 49.0, + "eval_loss": 4.384829521179199, + "eval_runtime": 49.1609, + "eval_samples_per_second": 49.674, + "eval_steps_per_second": 1.566, + "step": 3528 + }, + { + "epoch": 49.01397379912664, + "grad_norm": 0.005620039068162441, + "learning_rate": 3.3100087443634336e-05, + "loss": 3.671109676361084, + "step": 3529 + }, + { + "epoch": 49.02794759825328, + "grad_norm": 0.0053699444979429245, + "learning_rate": 3.301460085939133e-05, + "loss": 3.6248927116394043, + "step": 3530 + }, + { + "epoch": 49.041921397379916, + "grad_norm": 0.005503964144736528, + "learning_rate": 3.2930308284016734e-05, + "loss": 3.653135299682617, + "step": 3531 + }, + { + "epoch": 49.05589519650655, + "grad_norm": 0.00538219278678298, + "learning_rate": 3.284720978542336e-05, + "loss": 3.6291279792785645, + "step": 3532 + }, + { + "epoch": 49.069868995633186, + "grad_norm": 0.005521378945559263, + "learning_rate": 3.2765305430562e-05, + "loss": 3.6505486965179443, + "step": 3533 + }, + { + "epoch": 49.083842794759825, + "grad_norm": 0.005552167072892189, + "learning_rate": 3.268459528542179e-05, + "loss": 3.6507344245910645, + "step": 3534 + }, + { + "epoch": 49.09781659388646, + "grad_norm": 0.005520711652934551, + "learning_rate": 3.260507941502933e-05, + "loss": 3.63080096244812, + "step": 3535 + }, + { + "epoch": 49.1117903930131, + "grad_norm": 0.005630863830447197, + "learning_rate": 3.25267578834491e-05, + "loss": 3.667436361312866, + "step": 3536 + }, + { + "epoch": 49.12576419213974, + "grad_norm": 0.005421407055109739, + "learning_rate": 3.244963075378337e-05, + "loss": 3.6779494285583496, + "step": 3537 + }, + { + "epoch": 49.13973799126637, + "grad_norm": 0.005393749102950096, + "learning_rate": 3.237369808817209e-05, + "loss": 3.6451895236968994, + "step": 3538 + }, + { + "epoch": 49.15371179039301, + "grad_norm": 0.005522069521248341, + "learning_rate": 3.229895994779271e-05, + "loss": 3.670633554458618, + "step": 3539 + }, + { + "epoch": 49.16768558951965, + "grad_norm": 0.0055035874247550964, + "learning_rate": 3.222541639286068e-05, + "loss": 3.606834888458252, + "step": 3540 + }, + { + "epoch": 49.18165938864629, + "grad_norm": 0.005393242929130793, + "learning_rate": 3.215306748262843e-05, + "loss": 3.611767292022705, + "step": 3541 + }, + { + "epoch": 49.19563318777293, + "grad_norm": 0.005510237067937851, + "learning_rate": 3.2081913275386245e-05, + "loss": 3.6726202964782715, + "step": 3542 + }, + { + "epoch": 49.209606986899566, + "grad_norm": 0.005437937099486589, + "learning_rate": 3.201195382846191e-05, + "loss": 3.676589250564575, + "step": 3543 + }, + { + "epoch": 49.223580786026204, + "grad_norm": 0.0056598675437271595, + "learning_rate": 3.1943189198220366e-05, + "loss": 3.663947105407715, + "step": 3544 + }, + { + "epoch": 49.237554585152836, + "grad_norm": 0.005362293682992458, + "learning_rate": 3.1875619440064064e-05, + "loss": 3.6216440200805664, + "step": 3545 + }, + { + "epoch": 49.251528384279474, + "grad_norm": 0.005413597449660301, + "learning_rate": 3.1809244608432616e-05, + "loss": 3.6451244354248047, + "step": 3546 + }, + { + "epoch": 49.26550218340611, + "grad_norm": 0.00545527646318078, + "learning_rate": 3.174406475680331e-05, + "loss": 3.6419572830200195, + "step": 3547 + }, + { + "epoch": 49.27947598253275, + "grad_norm": 0.005393492989242077, + "learning_rate": 3.16800799376901e-05, + "loss": 3.6303319931030273, + "step": 3548 + }, + { + "epoch": 49.29344978165939, + "grad_norm": 0.005528928246349096, + "learning_rate": 3.1617290202644615e-05, + "loss": 3.631877899169922, + "step": 3549 + }, + { + "epoch": 49.30742358078603, + "grad_norm": 0.005303470883518457, + "learning_rate": 3.155569560225531e-05, + "loss": 3.6186366081237793, + "step": 3550 + }, + { + "epoch": 49.32139737991266, + "grad_norm": 0.005345781799405813, + "learning_rate": 3.1495296186147835e-05, + "loss": 3.636327028274536, + "step": 3551 + }, + { + "epoch": 49.3353711790393, + "grad_norm": 0.00533433398231864, + "learning_rate": 3.143609200298481e-05, + "loss": 3.612666606903076, + "step": 3552 + }, + { + "epoch": 49.34934497816594, + "grad_norm": 0.005468505900353193, + "learning_rate": 3.137808310046605e-05, + "loss": 3.6335525512695312, + "step": 3553 + }, + { + "epoch": 49.36331877729258, + "grad_norm": 0.005498886574059725, + "learning_rate": 3.132126952532838e-05, + "loss": 3.64517879486084, + "step": 3554 + }, + { + "epoch": 49.377292576419215, + "grad_norm": 0.005366173572838306, + "learning_rate": 3.1265651323345436e-05, + "loss": 3.6120004653930664, + "step": 3555 + }, + { + "epoch": 49.391266375545854, + "grad_norm": 0.005579676479101181, + "learning_rate": 3.121122853932773e-05, + "loss": 3.651233196258545, + "step": 3556 + }, + { + "epoch": 49.40524017467249, + "grad_norm": 0.005380513612180948, + "learning_rate": 3.115800121712275e-05, + "loss": 3.6802916526794434, + "step": 3557 + }, + { + "epoch": 49.419213973799124, + "grad_norm": 0.005381321534514427, + "learning_rate": 3.110596939961485e-05, + "loss": 3.606739044189453, + "step": 3558 + }, + { + "epoch": 49.43318777292576, + "grad_norm": 0.005590811837464571, + "learning_rate": 3.105513312872503e-05, + "loss": 3.6593852043151855, + "step": 3559 + }, + { + "epoch": 49.4471615720524, + "grad_norm": 0.005552604794502258, + "learning_rate": 3.100549244541116e-05, + "loss": 3.6589818000793457, + "step": 3560 + }, + { + "epoch": 49.46113537117904, + "grad_norm": 0.005272792186588049, + "learning_rate": 3.0957047389667774e-05, + "loss": 3.6500585079193115, + "step": 3561 + }, + { + "epoch": 49.47510917030568, + "grad_norm": 0.00557938776910305, + "learning_rate": 3.09097980005264e-05, + "loss": 3.6423044204711914, + "step": 3562 + }, + { + "epoch": 49.48908296943232, + "grad_norm": 0.005462596192955971, + "learning_rate": 3.0863744316054773e-05, + "loss": 3.6106159687042236, + "step": 3563 + }, + { + "epoch": 49.50305676855895, + "grad_norm": 0.005449238698929548, + "learning_rate": 3.081888637335776e-05, + "loss": 3.6468400955200195, + "step": 3564 + }, + { + "epoch": 49.51703056768559, + "grad_norm": 0.005447402596473694, + "learning_rate": 3.077522420857645e-05, + "loss": 3.624122142791748, + "step": 3565 + }, + { + "epoch": 49.531004366812226, + "grad_norm": 0.005388482473790646, + "learning_rate": 3.0732757856888576e-05, + "loss": 3.6392014026641846, + "step": 3566 + }, + { + "epoch": 49.544978165938865, + "grad_norm": 0.00541285565122962, + "learning_rate": 3.069148735250873e-05, + "loss": 3.654369354248047, + "step": 3567 + }, + { + "epoch": 49.5589519650655, + "grad_norm": 0.005579780787229538, + "learning_rate": 3.065141272868768e-05, + "loss": 3.6372079849243164, + "step": 3568 + }, + { + "epoch": 49.57292576419214, + "grad_norm": 0.005476080346852541, + "learning_rate": 3.061253401771287e-05, + "loss": 3.621183395385742, + "step": 3569 + }, + { + "epoch": 49.58689956331878, + "grad_norm": 0.005435108672827482, + "learning_rate": 3.057485125090808e-05, + "loss": 3.6125760078430176, + "step": 3570 + }, + { + "epoch": 49.60087336244541, + "grad_norm": 0.005412368569523096, + "learning_rate": 3.0538364458633627e-05, + "loss": 3.6276559829711914, + "step": 3571 + }, + { + "epoch": 49.61484716157205, + "grad_norm": 0.005578193813562393, + "learning_rate": 3.050307367028646e-05, + "loss": 3.672255516052246, + "step": 3572 + }, + { + "epoch": 49.62882096069869, + "grad_norm": 0.005397310480475426, + "learning_rate": 3.046897891429959e-05, + "loss": 3.6349916458129883, + "step": 3573 + }, + { + "epoch": 49.64279475982533, + "grad_norm": 0.005437100306153297, + "learning_rate": 3.0436080218142514e-05, + "loss": 3.641871929168701, + "step": 3574 + }, + { + "epoch": 49.65676855895197, + "grad_norm": 0.00533803878352046, + "learning_rate": 3.0404377608321075e-05, + "loss": 3.6466495990753174, + "step": 3575 + }, + { + "epoch": 49.670742358078606, + "grad_norm": 0.005386525299400091, + "learning_rate": 3.0373871110377803e-05, + "loss": 3.6554203033447266, + "step": 3576 + }, + { + "epoch": 49.68471615720524, + "grad_norm": 0.005472255405038595, + "learning_rate": 3.0344560748890904e-05, + "loss": 3.6272778511047363, + "step": 3577 + }, + { + "epoch": 49.698689956331876, + "grad_norm": 0.005554827395826578, + "learning_rate": 3.0316446547475262e-05, + "loss": 3.6589441299438477, + "step": 3578 + }, + { + "epoch": 49.712663755458514, + "grad_norm": 0.005507173947989941, + "learning_rate": 3.0289528528782108e-05, + "loss": 3.660010814666748, + "step": 3579 + }, + { + "epoch": 49.72663755458515, + "grad_norm": 0.005387808196246624, + "learning_rate": 3.0263806714498688e-05, + "loss": 3.6784188747406006, + "step": 3580 + }, + { + "epoch": 49.74061135371179, + "grad_norm": 0.005291743203997612, + "learning_rate": 3.0239281125348596e-05, + "loss": 3.665433168411255, + "step": 3581 + }, + { + "epoch": 49.75458515283843, + "grad_norm": 0.005448203068226576, + "learning_rate": 3.0215951781091777e-05, + "loss": 3.6186153888702393, + "step": 3582 + }, + { + "epoch": 49.76855895196506, + "grad_norm": 0.005570375360548496, + "learning_rate": 3.0193818700524177e-05, + "loss": 3.6409573554992676, + "step": 3583 + }, + { + "epoch": 49.7825327510917, + "grad_norm": 0.0053171273320913315, + "learning_rate": 3.0172881901477933e-05, + "loss": 3.627653121948242, + "step": 3584 + }, + { + "epoch": 49.79650655021834, + "grad_norm": 0.0054213316179811954, + "learning_rate": 3.0153141400821696e-05, + "loss": 3.6836981773376465, + "step": 3585 + }, + { + "epoch": 49.81048034934498, + "grad_norm": 0.005401757545769215, + "learning_rate": 3.0134597214459794e-05, + "loss": 3.665957450866699, + "step": 3586 + }, + { + "epoch": 49.82445414847162, + "grad_norm": 0.005455177277326584, + "learning_rate": 3.0117249357333078e-05, + "loss": 3.613741874694824, + "step": 3587 + }, + { + "epoch": 49.838427947598255, + "grad_norm": 0.005288220476359129, + "learning_rate": 3.0101097843418394e-05, + "loss": 3.66194486618042, + "step": 3588 + }, + { + "epoch": 49.852401746724894, + "grad_norm": 0.005295758601278067, + "learning_rate": 3.0086142685728467e-05, + "loss": 3.6889102458953857, + "step": 3589 + }, + { + "epoch": 49.866375545851525, + "grad_norm": 0.005356399342417717, + "learning_rate": 3.0072383896312843e-05, + "loss": 3.6496775150299072, + "step": 3590 + }, + { + "epoch": 49.880349344978164, + "grad_norm": 0.0054651242680847645, + "learning_rate": 3.0059821486256425e-05, + "loss": 3.6584572792053223, + "step": 3591 + }, + { + "epoch": 49.8943231441048, + "grad_norm": 0.005395952612161636, + "learning_rate": 3.004845546568064e-05, + "loss": 3.633230686187744, + "step": 3592 + }, + { + "epoch": 49.90829694323144, + "grad_norm": 0.005331660155206919, + "learning_rate": 3.003828584374276e-05, + "loss": 3.628465175628662, + "step": 3593 + }, + { + "epoch": 49.92227074235808, + "grad_norm": 0.0054406700655817986, + "learning_rate": 3.0029312628636385e-05, + "loss": 3.665616512298584, + "step": 3594 + }, + { + "epoch": 49.93624454148472, + "grad_norm": 0.005537047050893307, + "learning_rate": 3.0021535827590996e-05, + "loss": 3.650026321411133, + "step": 3595 + }, + { + "epoch": 49.95021834061135, + "grad_norm": 0.005354189779609442, + "learning_rate": 3.0014955446872235e-05, + "loss": 3.6486058235168457, + "step": 3596 + }, + { + "epoch": 49.96419213973799, + "grad_norm": 0.0054948581382632256, + "learning_rate": 3.000957149178194e-05, + "loss": 3.632627010345459, + "step": 3597 + }, + { + "epoch": 49.97816593886463, + "grad_norm": 0.00550198694691062, + "learning_rate": 3.000538396665778e-05, + "loss": 3.6374807357788086, + "step": 3598 + }, + { + "epoch": 49.992139737991266, + "grad_norm": 0.005545698571950197, + "learning_rate": 3.000239287487345e-05, + "loss": 3.6445484161376953, + "step": 3599 + }, + { + "epoch": 50.0, + "grad_norm": 0.006927069742232561, + "learning_rate": 3.0000598218838843e-05, + "loss": 3.6755237579345703, + "step": 3600 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.534806680666112e+18, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-3600/training_args.bin b/runs/baseline/checkpoint-3600/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-3600/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-72/chat_template.jinja b/runs/baseline/checkpoint-72/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-72/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-72/config.json b/runs/baseline/checkpoint-72/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-72/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-72/generation_config.json b/runs/baseline/checkpoint-72/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-72/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-72/model.safetensors b/runs/baseline/checkpoint-72/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9c5da32c04bbbd011788bcc0535c61e1d4fb4994 --- /dev/null +++ b/runs/baseline/checkpoint-72/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b68864cea94a3e73fe9caa2a0dc90451588ffa255ab33997e5376c85df693ab2 +size 583356232 diff --git a/runs/baseline/checkpoint-72/optimizer.pt b/runs/baseline/checkpoint-72/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..0d44ac376410c76eddf7dab317ac0077f2594c35 --- /dev/null +++ b/runs/baseline/checkpoint-72/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5f8e15426c5c1eb16331c156404fa8cd028829039d2caf56daa8739d001fe8c +size 1166825803 diff --git a/runs/baseline/checkpoint-72/rng_state_0.pth b/runs/baseline/checkpoint-72/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..3fc68e18ddaf65dfbdec55893d9a925ff5e43a18 --- /dev/null +++ b/runs/baseline/checkpoint-72/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4dd7671ce88d469c49c0530724ac76b2306574002d1ecd1ca9294e41621fd96a +size 14917 diff --git a/runs/baseline/checkpoint-72/rng_state_1.pth b/runs/baseline/checkpoint-72/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..79d9de29ae34b3b0c10ea8ba0348aeafb0c12226 --- /dev/null +++ b/runs/baseline/checkpoint-72/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3246ef1170ccca541a03b89ad6f20e01c51eb6834a2c2211c78c71c70f896879 +size 14917 diff --git a/runs/baseline/checkpoint-72/scheduler.pt b/runs/baseline/checkpoint-72/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..d33c2d0d0a3862ff1e2785405371ddce28072435 --- /dev/null +++ b/runs/baseline/checkpoint-72/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae8ce601b5ae4abd83f2c331115c5c2500c7c5c7f6bc8590a847682f194a23a4 +size 1465 diff --git a/runs/baseline/checkpoint-72/tokenizer.json b/runs/baseline/checkpoint-72/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-72/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-72/tokenizer_config.json b/runs/baseline/checkpoint-72/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-72/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-72/trainer_state.json b/runs/baseline/checkpoint-72/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..7afea72c7a477c337cded595f728ef51ea45e701 --- /dev/null +++ b/runs/baseline/checkpoint-72/trainer_state.json @@ -0,0 +1,538 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 72, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+16, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-72/training_args.bin b/runs/baseline/checkpoint-72/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-72/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265 diff --git a/runs/baseline/checkpoint-720/chat_template.jinja b/runs/baseline/checkpoint-720/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..699ff8df401fe4788525e9c1f9b86a99eadd6230 --- /dev/null +++ b/runs/baseline/checkpoint-720/chat_template.jinja @@ -0,0 +1,85 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0].role == 'system' %} + {{- messages[0].content + '\n\n' }} + {%- endif %} + {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0].role == 'system' %} + {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %} +{%- for message in messages[::-1] %} + {%- set index = (messages|length - 1) - loop.index0 %} + {%- if ns.multi_step_tool and message.role == "user" and not(message.content.startswith('') and message.content.endswith('')) %} + {%- set ns.multi_step_tool = false %} + {%- set ns.last_query_index = index %} + {%- endif %} +{%- endfor %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {%- set content = message.content %} + {%- set reasoning_content = '' %} + {%- if message.reasoning_content is defined and message.reasoning_content is not none %} + {%- set reasoning_content = message.reasoning_content %} + {%- else %} + {%- if '' in message.content %} + {%- set content = message.content.split('')[-1].lstrip('\n') %} + {%- set reasoning_content = message.content.split('')[0].rstrip('\n').split('')[-1].lstrip('\n') %} + {%- endif %} + {%- endif %} + {%- if loop.index0 > ns.last_query_index %} + {%- if loop.last or (not loop.last and reasoning_content) %} + {{- '<|im_start|>' + message.role + '\n\n' + reasoning_content.strip('\n') + '\n\n\n' + content.lstrip('\n') }} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- else %} + {{- '<|im_start|>' + message.role + '\n' + content }} + {%- endif %} + {%- if message.tool_calls %} + {%- for tool_call in message.tool_calls %} + {%- if (loop.first and content) or (not loop.first) %} + {{- '\n' }} + {%- endif %} + {%- if tool_call.function %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {%- if tool_call.arguments is string %} + {{- tool_call.arguments }} + {%- else %} + {{- tool_call.arguments | tojson }} + {%- endif %} + {{- '}\n' }} + {%- endfor %} + {%- endif %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} + {%- if enable_thinking is defined and enable_thinking is false %} + {{- '\n\n\n\n' }} + {%- endif %} +{%- endif %} \ No newline at end of file diff --git a/runs/baseline/checkpoint-720/config.json b/runs/baseline/checkpoint-720/config.json new file mode 100644 index 0000000000000000000000000000000000000000..44f7b99e220689bde520b89c11fdd83d1b5348de --- /dev/null +++ b/runs/baseline/checkpoint-720/config.json @@ -0,0 +1,32 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": null, + "dtype": "float32", + "eos_token_id": 151645, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 512, + "initializer_range": 0.02, + "intermediate_size": 1536, + "max_position_embeddings": 2048, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 4, + "num_hidden_layers": 20, + "num_key_value_heads": 4, + "pad_token_id": 151645, + "pretraining_tp": 1, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 10000.0, + "rope_type": "default" + }, + "tie_word_embeddings": true, + "transformers_version": "5.5.0", + "use_cache": false, + "vocab_size": 151671 +} diff --git a/runs/baseline/checkpoint-720/generation_config.json b/runs/baseline/checkpoint-720/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..f962c4bfc66159cdeb7ba836cbbd79365e9304f3 --- /dev/null +++ b/runs/baseline/checkpoint-720/generation_config.json @@ -0,0 +1,11 @@ +{ + "_from_model_config": true, + "eos_token_id": [ + 151645 + ], + "output_attentions": false, + "output_hidden_states": false, + "pad_token_id": 151645, + "transformers_version": "5.5.0", + "use_cache": true +} diff --git a/runs/baseline/checkpoint-720/model.safetensors b/runs/baseline/checkpoint-720/model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e76e4b365029a6aa0f0daa39af6bb2ac736fbb1c --- /dev/null +++ b/runs/baseline/checkpoint-720/model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91f481106cb2d8dcc4f7d57ef7ae377bb737fa8611a827eec3ed1e957fad3d2c +size 583356232 diff --git a/runs/baseline/checkpoint-720/optimizer.pt b/runs/baseline/checkpoint-720/optimizer.pt new file mode 100644 index 0000000000000000000000000000000000000000..b8c25eac628166256f178d45c0366ae78755dc5a --- /dev/null +++ b/runs/baseline/checkpoint-720/optimizer.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e60c4d7df3ee39416f5e537e1935b44bc5d1088ee3a74cdabef6b771df36e9d7 +size 1166825803 diff --git a/runs/baseline/checkpoint-720/rng_state_0.pth b/runs/baseline/checkpoint-720/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..64991ceae73e5d3389476a9b448c3cda96f73b10 --- /dev/null +++ b/runs/baseline/checkpoint-720/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed1b9484e9cec07c14c677aaf97ebf9dff6fa34fc12bae83c65f98c9d2c9ad0d +size 14917 diff --git a/runs/baseline/checkpoint-720/rng_state_1.pth b/runs/baseline/checkpoint-720/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..fee3b6b64b847dcce05678311e2696b568d3185a --- /dev/null +++ b/runs/baseline/checkpoint-720/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d30790dc4fc4171a789605cefdf649fd85c4d4afb59efd35ee6fea409dad4915 +size 14917 diff --git a/runs/baseline/checkpoint-720/scheduler.pt b/runs/baseline/checkpoint-720/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..454d8217902b1d6b2ad959de62cd253bdcce51c8 --- /dev/null +++ b/runs/baseline/checkpoint-720/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29fc245c65149711c0ae7f26d9977031181d64eae033b8ce0b653bc6512fd5e0 +size 1465 diff --git a/runs/baseline/checkpoint-720/tokenizer.json b/runs/baseline/checkpoint-720/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..b83d93986de8ecfc4343943be1c86e9532682c15 --- /dev/null +++ b/runs/baseline/checkpoint-720/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59b6776030505d5b441d0727cce137047df748ab15db6ba3a1bac93c123742fb +size 11424079 diff --git a/runs/baseline/checkpoint-720/tokenizer_config.json b/runs/baseline/checkpoint-720/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..b1d6bb391bb8fbc446de576db5220ec91c98a150 --- /dev/null +++ b/runs/baseline/checkpoint-720/tokenizer_config.json @@ -0,0 +1,18 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "extra_special_tokens": [ + "<|l2r_pred|>", + "<|r2l_pred|>" + ], + "is_local": false, + "model_max_length": 131072, + "pad_token": "<|im_end|>", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null +} diff --git a/runs/baseline/checkpoint-720/trainer_state.json b/runs/baseline/checkpoint-720/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..d7ceb5c73632a2a8723625795bfb0dfe919c03d1 --- /dev/null +++ b/runs/baseline/checkpoint-720/trainer_state.json @@ -0,0 +1,5146 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 10.0, + "eval_steps": 500, + "global_step": 720, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.013973799126637555, + "grad_norm": 0.12165174633264542, + "learning_rate": 0.0, + "loss": 12.027063369750977, + "step": 1 + }, + { + "epoch": 0.02794759825327511, + "grad_norm": 0.12109383195638657, + "learning_rate": 3e-05, + "loss": 12.028635025024414, + "step": 2 + }, + { + "epoch": 0.04192139737991266, + "grad_norm": 0.12690064311027527, + "learning_rate": 6e-05, + "loss": 11.894524574279785, + "step": 3 + }, + { + "epoch": 0.05589519650655022, + "grad_norm": 0.1537286937236786, + "learning_rate": 8.999999999999999e-05, + "loss": 11.672262191772461, + "step": 4 + }, + { + "epoch": 0.06986899563318777, + "grad_norm": 0.11778244376182556, + "learning_rate": 0.00012, + "loss": 11.433671951293945, + "step": 5 + }, + { + "epoch": 0.08384279475982533, + "grad_norm": 0.10874935984611511, + "learning_rate": 0.00015000000000000001, + "loss": 11.270309448242188, + "step": 6 + }, + { + "epoch": 0.09781659388646288, + "grad_norm": 0.10641921311616898, + "learning_rate": 0.00017999999999999998, + "loss": 11.155481338500977, + "step": 7 + }, + { + "epoch": 0.11179039301310044, + "grad_norm": 0.1048506498336792, + "learning_rate": 0.00021000000000000004, + "loss": 11.066675186157227, + "step": 8 + }, + { + "epoch": 0.125764192139738, + "grad_norm": 0.10546834021806717, + "learning_rate": 0.00024, + "loss": 10.952295303344727, + "step": 9 + }, + { + "epoch": 0.13973799126637554, + "grad_norm": 0.10614268481731415, + "learning_rate": 0.00027, + "loss": 10.811386108398438, + "step": 10 + }, + { + "epoch": 0.1537117903930131, + "grad_norm": 0.10529010742902756, + "learning_rate": 0.00030000000000000003, + "loss": 10.657712936401367, + "step": 11 + }, + { + "epoch": 0.16768558951965065, + "grad_norm": 0.1051444336771965, + "learning_rate": 0.00033, + "loss": 10.47591781616211, + "step": 12 + }, + { + "epoch": 0.18165938864628822, + "grad_norm": 0.10420846194028854, + "learning_rate": 0.00035999999999999997, + "loss": 10.280141830444336, + "step": 13 + }, + { + "epoch": 0.19563318777292577, + "grad_norm": 0.10323625057935715, + "learning_rate": 0.00039000000000000005, + "loss": 10.076366424560547, + "step": 14 + }, + { + "epoch": 0.2096069868995633, + "grad_norm": 0.10255401581525803, + "learning_rate": 0.00042000000000000007, + "loss": 9.856939315795898, + "step": 15 + }, + { + "epoch": 0.22358078602620088, + "grad_norm": 0.10151027888059616, + "learning_rate": 0.00045, + "loss": 9.632134437561035, + "step": 16 + }, + { + "epoch": 0.23755458515283842, + "grad_norm": 0.1006326898932457, + "learning_rate": 0.00048, + "loss": 9.401637077331543, + "step": 17 + }, + { + "epoch": 0.251528384279476, + "grad_norm": 0.09820562601089478, + "learning_rate": 0.00051, + "loss": 9.177400588989258, + "step": 18 + }, + { + "epoch": 0.26550218340611353, + "grad_norm": 0.0956626906991005, + "learning_rate": 0.00054, + "loss": 8.95518684387207, + "step": 19 + }, + { + "epoch": 0.2794759825327511, + "grad_norm": 0.0922091156244278, + "learning_rate": 0.00057, + "loss": 8.75066089630127, + "step": 20 + }, + { + "epoch": 0.2934497816593886, + "grad_norm": 0.08593932539224625, + "learning_rate": 0.0006000000000000001, + "loss": 8.557976722717285, + "step": 21 + }, + { + "epoch": 0.3074235807860262, + "grad_norm": 0.08014766871929169, + "learning_rate": 0.00063, + "loss": 8.360604286193848, + "step": 22 + }, + { + "epoch": 0.32139737991266376, + "grad_norm": 0.07381244748830795, + "learning_rate": 0.00066, + "loss": 8.210375785827637, + "step": 23 + }, + { + "epoch": 0.3353711790393013, + "grad_norm": 0.06455052644014359, + "learning_rate": 0.0006900000000000001, + "loss": 8.07168960571289, + "step": 24 + }, + { + "epoch": 0.34934497816593885, + "grad_norm": 0.054115474224090576, + "learning_rate": 0.0007199999999999999, + "loss": 7.960069179534912, + "step": 25 + }, + { + "epoch": 0.36331877729257644, + "grad_norm": 0.043575264513492584, + "learning_rate": 0.00075, + "loss": 7.850685119628906, + "step": 26 + }, + { + "epoch": 0.377292576419214, + "grad_norm": 0.02976679615676403, + "learning_rate": 0.0007800000000000001, + "loss": 7.824377059936523, + "step": 27 + }, + { + "epoch": 0.39126637554585153, + "grad_norm": 0.020552165806293488, + "learning_rate": 0.0008100000000000001, + "loss": 7.7695417404174805, + "step": 28 + }, + { + "epoch": 0.4052401746724891, + "grad_norm": 0.018530743196606636, + "learning_rate": 0.0008400000000000001, + "loss": 7.782247543334961, + "step": 29 + }, + { + "epoch": 0.4192139737991266, + "grad_norm": 0.02716820128262043, + "learning_rate": 0.00087, + "loss": 7.754945278167725, + "step": 30 + }, + { + "epoch": 0.4331877729257642, + "grad_norm": 0.03638409823179245, + "learning_rate": 0.0009, + "loss": 7.799381256103516, + "step": 31 + }, + { + "epoch": 0.44716157205240176, + "grad_norm": 0.041579652577638626, + "learning_rate": 0.00093, + "loss": 7.776629447937012, + "step": 32 + }, + { + "epoch": 0.4611353711790393, + "grad_norm": 0.0304960235953331, + "learning_rate": 0.00096, + "loss": 7.749439239501953, + "step": 33 + }, + { + "epoch": 0.47510917030567684, + "grad_norm": 0.053728822618722916, + "learning_rate": 0.00099, + "loss": 7.745059490203857, + "step": 34 + }, + { + "epoch": 0.4890829694323144, + "grad_norm": 0.036322493106126785, + "learning_rate": 0.00102, + "loss": 7.752074718475342, + "step": 35 + }, + { + "epoch": 0.503056768558952, + "grad_norm": 0.03232959285378456, + "learning_rate": 0.00105, + "loss": 7.704257011413574, + "step": 36 + }, + { + "epoch": 0.5170305676855895, + "grad_norm": 0.03246184438467026, + "learning_rate": 0.00108, + "loss": 7.667045593261719, + "step": 37 + }, + { + "epoch": 0.5310043668122271, + "grad_norm": 0.036655545234680176, + "learning_rate": 0.00111, + "loss": 7.623445987701416, + "step": 38 + }, + { + "epoch": 0.5449781659388646, + "grad_norm": 0.021042725071310997, + "learning_rate": 0.00114, + "loss": 7.582821846008301, + "step": 39 + }, + { + "epoch": 0.5589519650655022, + "grad_norm": 0.022982975468039513, + "learning_rate": 0.00117, + "loss": 7.58036994934082, + "step": 40 + }, + { + "epoch": 0.5729257641921397, + "grad_norm": 0.024781864136457443, + "learning_rate": 0.0012000000000000001, + "loss": 7.5903000831604, + "step": 41 + }, + { + "epoch": 0.5868995633187772, + "grad_norm": 0.03667089343070984, + "learning_rate": 0.00123, + "loss": 7.588233947753906, + "step": 42 + }, + { + "epoch": 0.6008733624454149, + "grad_norm": 0.04083017632365227, + "learning_rate": 0.00126, + "loss": 7.548343658447266, + "step": 43 + }, + { + "epoch": 0.6148471615720524, + "grad_norm": 0.02585385926067829, + "learning_rate": 0.00129, + "loss": 7.604833602905273, + "step": 44 + }, + { + "epoch": 0.62882096069869, + "grad_norm": 0.018793931230902672, + "learning_rate": 0.00132, + "loss": 7.56705379486084, + "step": 45 + }, + { + "epoch": 0.6427947598253275, + "grad_norm": 0.025538936257362366, + "learning_rate": 0.00135, + "loss": 7.562107086181641, + "step": 46 + }, + { + "epoch": 0.6567685589519651, + "grad_norm": 0.02819318138062954, + "learning_rate": 0.0013800000000000002, + "loss": 7.517213821411133, + "step": 47 + }, + { + "epoch": 0.6707423580786026, + "grad_norm": 0.016239255666732788, + "learning_rate": 0.00141, + "loss": 7.4780755043029785, + "step": 48 + }, + { + "epoch": 0.6847161572052402, + "grad_norm": 0.024470701813697815, + "learning_rate": 0.0014399999999999999, + "loss": 7.505887985229492, + "step": 49 + }, + { + "epoch": 0.6986899563318777, + "grad_norm": 0.015499847941100597, + "learning_rate": 0.00147, + "loss": 7.484469890594482, + "step": 50 + }, + { + "epoch": 0.7126637554585152, + "grad_norm": 0.017785200849175453, + "learning_rate": 0.0015, + "loss": 7.481822967529297, + "step": 51 + }, + { + "epoch": 0.7266375545851529, + "grad_norm": 0.015952805057168007, + "learning_rate": 0.0015300000000000001, + "loss": 7.484804153442383, + "step": 52 + }, + { + "epoch": 0.7406113537117904, + "grad_norm": 0.015771260485053062, + "learning_rate": 0.0015600000000000002, + "loss": 7.4495439529418945, + "step": 53 + }, + { + "epoch": 0.754585152838428, + "grad_norm": 0.024370530620217323, + "learning_rate": 0.00159, + "loss": 7.471208572387695, + "step": 54 + }, + { + "epoch": 0.7685589519650655, + "grad_norm": 0.018173737451434135, + "learning_rate": 0.0016200000000000001, + "loss": 7.438447952270508, + "step": 55 + }, + { + "epoch": 0.7825327510917031, + "grad_norm": 0.020564094185829163, + "learning_rate": 0.0016500000000000002, + "loss": 7.437922954559326, + "step": 56 + }, + { + "epoch": 0.7965065502183406, + "grad_norm": 0.016154944896697998, + "learning_rate": 0.0016800000000000003, + "loss": 7.429378032684326, + "step": 57 + }, + { + "epoch": 0.8104803493449781, + "grad_norm": 0.016817007213830948, + "learning_rate": 0.00171, + "loss": 7.406428337097168, + "step": 58 + }, + { + "epoch": 0.8244541484716157, + "grad_norm": 0.03460947796702385, + "learning_rate": 0.00174, + "loss": 7.393331527709961, + "step": 59 + }, + { + "epoch": 0.8384279475982532, + "grad_norm": 0.15410488843917847, + "learning_rate": 0.0017699999999999999, + "loss": 7.645856857299805, + "step": 60 + }, + { + "epoch": 0.8524017467248908, + "grad_norm": 0.07443148642778397, + "learning_rate": 0.0018, + "loss": 7.549508094787598, + "step": 61 + }, + { + "epoch": 0.8663755458515284, + "grad_norm": 0.0507572740316391, + "learning_rate": 0.00183, + "loss": 7.517921447753906, + "step": 62 + }, + { + "epoch": 0.880349344978166, + "grad_norm": 0.09682746231555939, + "learning_rate": 0.00186, + "loss": 7.543978691101074, + "step": 63 + }, + { + "epoch": 0.8943231441048035, + "grad_norm": 0.2571517527103424, + "learning_rate": 0.00189, + "loss": 7.768949508666992, + "step": 64 + }, + { + "epoch": 0.9082969432314411, + "grad_norm": 0.0317784883081913, + "learning_rate": 0.00192, + "loss": 7.509063720703125, + "step": 65 + }, + { + "epoch": 0.9222707423580786, + "grad_norm": 0.0353970006108284, + "learning_rate": 0.0019500000000000001, + "loss": 7.522322654724121, + "step": 66 + }, + { + "epoch": 0.9362445414847161, + "grad_norm": 0.03663994371891022, + "learning_rate": 0.00198, + "loss": 7.504723072052002, + "step": 67 + }, + { + "epoch": 0.9502183406113537, + "grad_norm": 0.025481393560767174, + "learning_rate": 0.00201, + "loss": 7.452597141265869, + "step": 68 + }, + { + "epoch": 0.9641921397379912, + "grad_norm": 0.03099915198981762, + "learning_rate": 0.00204, + "loss": 7.455441474914551, + "step": 69 + }, + { + "epoch": 0.9781659388646288, + "grad_norm": 0.022921698167920113, + "learning_rate": 0.00207, + "loss": 7.4454827308654785, + "step": 70 + }, + { + "epoch": 0.9921397379912664, + "grad_norm": 0.032388731837272644, + "learning_rate": 0.0021, + "loss": 7.467103958129883, + "step": 71 + }, + { + "epoch": 1.0, + "grad_norm": 0.022428609430789948, + "learning_rate": 0.00213, + "loss": 7.467540264129639, + "step": 72 + }, + { + "epoch": 1.0, + "eval_loss": 7.469719409942627, + "eval_runtime": 50.9714, + "eval_samples_per_second": 47.909, + "eval_steps_per_second": 1.511, + "step": 72 + }, + { + "epoch": 1.0139737991266375, + "grad_norm": 0.017480911687016487, + "learning_rate": 0.00216, + "loss": 7.455810546875, + "step": 73 + }, + { + "epoch": 1.027947598253275, + "grad_norm": 0.016675131395459175, + "learning_rate": 0.00219, + "loss": 7.474848747253418, + "step": 74 + }, + { + "epoch": 1.0419213973799126, + "grad_norm": 0.017450228333473206, + "learning_rate": 0.00222, + "loss": 7.442611217498779, + "step": 75 + }, + { + "epoch": 1.0558951965065502, + "grad_norm": 0.016823675483465195, + "learning_rate": 0.0022500000000000003, + "loss": 7.454628944396973, + "step": 76 + }, + { + "epoch": 1.0698689956331877, + "grad_norm": 0.016717543825507164, + "learning_rate": 0.00228, + "loss": 7.425317764282227, + "step": 77 + }, + { + "epoch": 1.0838427947598253, + "grad_norm": 0.01644429750740528, + "learning_rate": 0.00231, + "loss": 7.4102864265441895, + "step": 78 + }, + { + "epoch": 1.0978165938864628, + "grad_norm": 0.013104400597512722, + "learning_rate": 0.00234, + "loss": 7.4559173583984375, + "step": 79 + }, + { + "epoch": 1.1117903930131003, + "grad_norm": 0.01175684668123722, + "learning_rate": 0.00237, + "loss": 7.438898086547852, + "step": 80 + }, + { + "epoch": 1.125764192139738, + "grad_norm": 0.013736684806644917, + "learning_rate": 0.0024000000000000002, + "loss": 7.449005603790283, + "step": 81 + }, + { + "epoch": 1.1397379912663754, + "grad_norm": 0.01176756713539362, + "learning_rate": 0.0024300000000000003, + "loss": 7.428421974182129, + "step": 82 + }, + { + "epoch": 1.1537117903930132, + "grad_norm": 0.011860918253660202, + "learning_rate": 0.00246, + "loss": 7.411491394042969, + "step": 83 + }, + { + "epoch": 1.1676855895196507, + "grad_norm": 0.010093477554619312, + "learning_rate": 0.00249, + "loss": 7.39311408996582, + "step": 84 + }, + { + "epoch": 1.1816593886462883, + "grad_norm": 0.009453288279473782, + "learning_rate": 0.00252, + "loss": 7.387656211853027, + "step": 85 + }, + { + "epoch": 1.1956331877729258, + "grad_norm": 0.010311692021787167, + "learning_rate": 0.00255, + "loss": 7.401076316833496, + "step": 86 + }, + { + "epoch": 1.2096069868995634, + "grad_norm": 0.010053195990622044, + "learning_rate": 0.00258, + "loss": 7.39655876159668, + "step": 87 + }, + { + "epoch": 1.223580786026201, + "grad_norm": 0.009758037514984608, + "learning_rate": 0.00261, + "loss": 7.423122882843018, + "step": 88 + }, + { + "epoch": 1.2375545851528384, + "grad_norm": 0.008394946344196796, + "learning_rate": 0.00264, + "loss": 7.383601188659668, + "step": 89 + }, + { + "epoch": 1.251528384279476, + "grad_norm": 0.008457187563180923, + "learning_rate": 0.00267, + "loss": 7.400934219360352, + "step": 90 + }, + { + "epoch": 1.2655021834061135, + "grad_norm": 0.010601334273815155, + "learning_rate": 0.0027, + "loss": 7.400241851806641, + "step": 91 + }, + { + "epoch": 1.279475982532751, + "grad_norm": 0.009009373374283314, + "learning_rate": 0.0027300000000000002, + "loss": 7.377348899841309, + "step": 92 + }, + { + "epoch": 1.2934497816593886, + "grad_norm": 0.00737508200109005, + "learning_rate": 0.0027600000000000003, + "loss": 7.385881423950195, + "step": 93 + }, + { + "epoch": 1.3074235807860262, + "grad_norm": 0.0075711836107075214, + "learning_rate": 0.0027900000000000004, + "loss": 7.348570823669434, + "step": 94 + }, + { + "epoch": 1.3213973799126637, + "grad_norm": 0.00846901722252369, + "learning_rate": 0.00282, + "loss": 7.364287853240967, + "step": 95 + }, + { + "epoch": 1.3353711790393012, + "grad_norm": 0.007160635199397802, + "learning_rate": 0.00285, + "loss": 7.3881635665893555, + "step": 96 + }, + { + "epoch": 1.3493449781659388, + "grad_norm": 0.00682067358866334, + "learning_rate": 0.0028799999999999997, + "loss": 7.348663330078125, + "step": 97 + }, + { + "epoch": 1.3633187772925766, + "grad_norm": 0.009108101949095726, + "learning_rate": 0.00291, + "loss": 7.356302261352539, + "step": 98 + }, + { + "epoch": 1.3772925764192139, + "grad_norm": 0.015828927978873253, + "learning_rate": 0.00294, + "loss": 7.347576141357422, + "step": 99 + }, + { + "epoch": 1.3912663755458516, + "grad_norm": 0.009683096781373024, + "learning_rate": 0.00297, + "loss": 7.323967933654785, + "step": 100 + }, + { + "epoch": 1.405240174672489, + "grad_norm": 0.010833270847797394, + "learning_rate": 0.003, + "loss": 7.3139495849609375, + "step": 101 + }, + { + "epoch": 1.4192139737991267, + "grad_norm": 0.014905400574207306, + "learning_rate": 0.002999999401781161, + "loss": 7.302030086517334, + "step": 102 + }, + { + "epoch": 1.4331877729257643, + "grad_norm": 0.023674853146076202, + "learning_rate": 0.002999997607125127, + "loss": 7.311811923980713, + "step": 103 + }, + { + "epoch": 1.4471615720524018, + "grad_norm": 0.050460800528526306, + "learning_rate": 0.002999994616033342, + "loss": 7.41972541809082, + "step": 104 + }, + { + "epoch": 1.4611353711790394, + "grad_norm": 0.02193174883723259, + "learning_rate": 0.002999990428508218, + "loss": 7.294468879699707, + "step": 105 + }, + { + "epoch": 1.475109170305677, + "grad_norm": 0.036261629313230515, + "learning_rate": 0.0029999850445531283, + "loss": 7.295568466186523, + "step": 106 + }, + { + "epoch": 1.4890829694323144, + "grad_norm": 0.02734929881989956, + "learning_rate": 0.002999978464172409, + "loss": 7.309239864349365, + "step": 107 + }, + { + "epoch": 1.503056768558952, + "grad_norm": 0.017884107306599617, + "learning_rate": 0.002999970687371363, + "loss": 7.2427592277526855, + "step": 108 + }, + { + "epoch": 1.5170305676855895, + "grad_norm": 0.01700308918952942, + "learning_rate": 0.0029999617141562573, + "loss": 7.241730213165283, + "step": 109 + }, + { + "epoch": 1.531004366812227, + "grad_norm": 0.0209935512393713, + "learning_rate": 0.0029999515445343194, + "loss": 7.262085437774658, + "step": 110 + }, + { + "epoch": 1.5449781659388646, + "grad_norm": 0.023207105696201324, + "learning_rate": 0.0029999401785137437, + "loss": 7.25450325012207, + "step": 111 + }, + { + "epoch": 1.5589519650655022, + "grad_norm": 0.01907353848218918, + "learning_rate": 0.002999927616103687, + "loss": 7.23149299621582, + "step": 112 + }, + { + "epoch": 1.5729257641921397, + "grad_norm": 0.016149666160345078, + "learning_rate": 0.0029999138573142715, + "loss": 7.210232734680176, + "step": 113 + }, + { + "epoch": 1.5868995633187772, + "grad_norm": 0.017513928934931755, + "learning_rate": 0.002999898902156582, + "loss": 7.218451976776123, + "step": 114 + }, + { + "epoch": 1.600873362445415, + "grad_norm": 0.015002547763288021, + "learning_rate": 0.002999882750642667, + "loss": 7.169316291809082, + "step": 115 + }, + { + "epoch": 1.6148471615720523, + "grad_norm": 0.014233081601560116, + "learning_rate": 0.0029998654027855406, + "loss": 7.194825172424316, + "step": 116 + }, + { + "epoch": 1.62882096069869, + "grad_norm": 0.01283260714262724, + "learning_rate": 0.0029998468585991783, + "loss": 7.15669059753418, + "step": 117 + }, + { + "epoch": 1.6427947598253274, + "grad_norm": 0.014416799880564213, + "learning_rate": 0.0029998271180985223, + "loss": 7.143369674682617, + "step": 118 + }, + { + "epoch": 1.6567685589519652, + "grad_norm": 0.014337058179080486, + "learning_rate": 0.0029998061812994757, + "loss": 7.127734184265137, + "step": 119 + }, + { + "epoch": 1.6707423580786025, + "grad_norm": 0.016406210139393806, + "learning_rate": 0.002999784048218908, + "loss": 7.086504936218262, + "step": 120 + }, + { + "epoch": 1.6847161572052403, + "grad_norm": 0.01680421642959118, + "learning_rate": 0.002999760718874652, + "loss": 7.089420318603516, + "step": 121 + }, + { + "epoch": 1.6986899563318776, + "grad_norm": 0.010665548034012318, + "learning_rate": 0.0029997361932855016, + "loss": 7.102463722229004, + "step": 122 + }, + { + "epoch": 1.7126637554585153, + "grad_norm": 0.013479137793183327, + "learning_rate": 0.002999710471471218, + "loss": 7.079098701477051, + "step": 123 + }, + { + "epoch": 1.726637554585153, + "grad_norm": 0.012408486567437649, + "learning_rate": 0.0029996835534525246, + "loss": 7.0790114402771, + "step": 124 + }, + { + "epoch": 1.7406113537117904, + "grad_norm": 0.012035196647047997, + "learning_rate": 0.0029996554392511094, + "loss": 7.058756351470947, + "step": 125 + }, + { + "epoch": 1.754585152838428, + "grad_norm": 0.00899985246360302, + "learning_rate": 0.0029996261288896225, + "loss": 7.0739240646362305, + "step": 126 + }, + { + "epoch": 1.7685589519650655, + "grad_norm": 0.01072906143963337, + "learning_rate": 0.002999595622391679, + "loss": 7.085537910461426, + "step": 127 + }, + { + "epoch": 1.782532751091703, + "grad_norm": 0.01047342736274004, + "learning_rate": 0.0029995639197818577, + "loss": 7.060049057006836, + "step": 128 + }, + { + "epoch": 1.7965065502183406, + "grad_norm": 0.02342618815600872, + "learning_rate": 0.0029995310210857006, + "loss": 7.039482116699219, + "step": 129 + }, + { + "epoch": 1.8104803493449781, + "grad_norm": 0.011256859637796879, + "learning_rate": 0.0029994969263297134, + "loss": 7.040412902832031, + "step": 130 + }, + { + "epoch": 1.8244541484716157, + "grad_norm": 0.008407019078731537, + "learning_rate": 0.0029994616355413666, + "loss": 7.0364274978637695, + "step": 131 + }, + { + "epoch": 1.8384279475982532, + "grad_norm": 0.011006148532032967, + "learning_rate": 0.0029994251487490916, + "loss": 6.991468906402588, + "step": 132 + }, + { + "epoch": 1.8524017467248908, + "grad_norm": 0.009103255346417427, + "learning_rate": 0.0029993874659822875, + "loss": 7.001059532165527, + "step": 133 + }, + { + "epoch": 1.8663755458515285, + "grad_norm": 0.007875788025557995, + "learning_rate": 0.0029993485872713125, + "loss": 6.9974260330200195, + "step": 134 + }, + { + "epoch": 1.8803493449781659, + "grad_norm": 0.007872611284255981, + "learning_rate": 0.0029993085126474913, + "loss": 6.980216979980469, + "step": 135 + }, + { + "epoch": 1.8943231441048036, + "grad_norm": 0.010071957483887672, + "learning_rate": 0.0029992672421431115, + "loss": 6.97799015045166, + "step": 136 + }, + { + "epoch": 1.908296943231441, + "grad_norm": 0.01209618803113699, + "learning_rate": 0.0029992247757914237, + "loss": 6.97202205657959, + "step": 137 + }, + { + "epoch": 1.9222707423580787, + "grad_norm": 0.014884858392179012, + "learning_rate": 0.0029991811136266425, + "loss": 6.976570129394531, + "step": 138 + }, + { + "epoch": 1.936244541484716, + "grad_norm": 0.014953463338315487, + "learning_rate": 0.0029991362556839453, + "loss": 6.971050262451172, + "step": 139 + }, + { + "epoch": 1.9502183406113538, + "grad_norm": 0.008371063508093357, + "learning_rate": 0.002999090201999474, + "loss": 6.955556869506836, + "step": 140 + }, + { + "epoch": 1.9641921397379911, + "grad_norm": 0.011926881037652493, + "learning_rate": 0.002999042952610332, + "loss": 6.943947792053223, + "step": 141 + }, + { + "epoch": 1.9781659388646289, + "grad_norm": 0.013422893360257149, + "learning_rate": 0.002998994507554589, + "loss": 6.910022735595703, + "step": 142 + }, + { + "epoch": 1.9921397379912664, + "grad_norm": 0.009672224521636963, + "learning_rate": 0.002998944866871275, + "loss": 6.8994855880737305, + "step": 143 + }, + { + "epoch": 2.0, + "grad_norm": 0.012590581551194191, + "learning_rate": 0.0029988940306003853, + "loss": 6.88709020614624, + "step": 144 + }, + { + "epoch": 2.0, + "eval_loss": 6.945146083831787, + "eval_runtime": 51.5801, + "eval_samples_per_second": 47.344, + "eval_steps_per_second": 1.493, + "step": 144 + }, + { + "epoch": 2.0139737991266378, + "grad_norm": 0.016554836183786392, + "learning_rate": 0.002998841998782877, + "loss": 6.916609764099121, + "step": 145 + }, + { + "epoch": 2.027947598253275, + "grad_norm": 0.011085660196840763, + "learning_rate": 0.0029987887714606723, + "loss": 6.884910583496094, + "step": 146 + }, + { + "epoch": 2.041921397379913, + "grad_norm": 0.008596168830990791, + "learning_rate": 0.0029987343486766547, + "loss": 6.900066375732422, + "step": 147 + }, + { + "epoch": 2.05589519650655, + "grad_norm": 0.014487302862107754, + "learning_rate": 0.0029986787304746715, + "loss": 6.8693342208862305, + "step": 148 + }, + { + "epoch": 2.069868995633188, + "grad_norm": 0.015636997297406197, + "learning_rate": 0.002998621916899534, + "loss": 6.903670787811279, + "step": 149 + }, + { + "epoch": 2.0838427947598253, + "grad_norm": 0.013275462202727795, + "learning_rate": 0.0029985639079970153, + "loss": 6.889361381530762, + "step": 150 + }, + { + "epoch": 2.097816593886463, + "grad_norm": 0.009044744074344635, + "learning_rate": 0.002998504703813852, + "loss": 6.8549981117248535, + "step": 151 + }, + { + "epoch": 2.1117903930131003, + "grad_norm": 0.011327696964144707, + "learning_rate": 0.002998444304397745, + "loss": 6.856151103973389, + "step": 152 + }, + { + "epoch": 2.125764192139738, + "grad_norm": 0.01310331467539072, + "learning_rate": 0.0029983827097973554, + "loss": 6.834944725036621, + "step": 153 + }, + { + "epoch": 2.1397379912663754, + "grad_norm": 0.010003753937780857, + "learning_rate": 0.00299831992006231, + "loss": 6.825613975524902, + "step": 154 + }, + { + "epoch": 2.153711790393013, + "grad_norm": 0.009928115643560886, + "learning_rate": 0.0029982559352431967, + "loss": 6.828935146331787, + "step": 155 + }, + { + "epoch": 2.1676855895196505, + "grad_norm": 0.012302150949835777, + "learning_rate": 0.0029981907553915674, + "loss": 6.821207046508789, + "step": 156 + }, + { + "epoch": 2.1816593886462883, + "grad_norm": 0.020555168390274048, + "learning_rate": 0.002998124380559936, + "loss": 6.8184895515441895, + "step": 157 + }, + { + "epoch": 2.1956331877729256, + "grad_norm": 0.011417638510465622, + "learning_rate": 0.0029980568108017797, + "loss": 6.788754463195801, + "step": 158 + }, + { + "epoch": 2.2096069868995634, + "grad_norm": 0.010772978886961937, + "learning_rate": 0.002997988046171538, + "loss": 6.7861433029174805, + "step": 159 + }, + { + "epoch": 2.2235807860262007, + "grad_norm": 0.012753859162330627, + "learning_rate": 0.002997918086724614, + "loss": 6.788776397705078, + "step": 160 + }, + { + "epoch": 2.2375545851528384, + "grad_norm": 0.011327114887535572, + "learning_rate": 0.0029978469325173717, + "loss": 6.779160022735596, + "step": 161 + }, + { + "epoch": 2.251528384279476, + "grad_norm": 0.01235753484070301, + "learning_rate": 0.002997774583607139, + "loss": 6.799554824829102, + "step": 162 + }, + { + "epoch": 2.2655021834061135, + "grad_norm": 0.01417156495153904, + "learning_rate": 0.0029977010400522074, + "loss": 6.77957820892334, + "step": 163 + }, + { + "epoch": 2.279475982532751, + "grad_norm": 0.021321553736925125, + "learning_rate": 0.002997626301911828, + "loss": 6.785632133483887, + "step": 164 + }, + { + "epoch": 2.2934497816593886, + "grad_norm": 0.03614472970366478, + "learning_rate": 0.002997550369246217, + "loss": 6.836414337158203, + "step": 165 + }, + { + "epoch": 2.3074235807860264, + "grad_norm": 0.03934052586555481, + "learning_rate": 0.002997473242116551, + "loss": 6.854659557342529, + "step": 166 + }, + { + "epoch": 2.3213973799126637, + "grad_norm": 0.016240060329437256, + "learning_rate": 0.0029973949205849707, + "loss": 6.789690971374512, + "step": 167 + }, + { + "epoch": 2.3353711790393015, + "grad_norm": 0.031164905056357384, + "learning_rate": 0.0029973154047145782, + "loss": 6.8363776206970215, + "step": 168 + }, + { + "epoch": 2.349344978165939, + "grad_norm": 0.013051844201982021, + "learning_rate": 0.002997234694569438, + "loss": 6.803576469421387, + "step": 169 + }, + { + "epoch": 2.3633187772925766, + "grad_norm": 0.022611448541283607, + "learning_rate": 0.002997152790214577, + "loss": 6.762107849121094, + "step": 170 + }, + { + "epoch": 2.377292576419214, + "grad_norm": 0.017812907695770264, + "learning_rate": 0.0029970696917159833, + "loss": 6.776218414306641, + "step": 171 + }, + { + "epoch": 2.3912663755458516, + "grad_norm": 0.015453332103788853, + "learning_rate": 0.0029969853991406088, + "loss": 6.742618083953857, + "step": 172 + }, + { + "epoch": 2.405240174672489, + "grad_norm": 0.020630091428756714, + "learning_rate": 0.0029968999125563655, + "loss": 6.765475749969482, + "step": 173 + }, + { + "epoch": 2.4192139737991267, + "grad_norm": 0.014672341756522655, + "learning_rate": 0.0029968132320321297, + "loss": 6.754009246826172, + "step": 174 + }, + { + "epoch": 2.433187772925764, + "grad_norm": 0.016540220007300377, + "learning_rate": 0.0029967253576377376, + "loss": 6.734752655029297, + "step": 175 + }, + { + "epoch": 2.447161572052402, + "grad_norm": 0.013495649211108685, + "learning_rate": 0.0029966362894439874, + "loss": 6.752676010131836, + "step": 176 + }, + { + "epoch": 2.461135371179039, + "grad_norm": 0.015936989337205887, + "learning_rate": 0.0029965460275226405, + "loss": 6.725708961486816, + "step": 177 + }, + { + "epoch": 2.475109170305677, + "grad_norm": 0.010431243106722832, + "learning_rate": 0.0029964545719464194, + "loss": 6.698376655578613, + "step": 178 + }, + { + "epoch": 2.489082969432314, + "grad_norm": 0.01669321395456791, + "learning_rate": 0.0029963619227890083, + "loss": 6.711770534515381, + "step": 179 + }, + { + "epoch": 2.503056768558952, + "grad_norm": 0.013308997265994549, + "learning_rate": 0.0029962680801250514, + "loss": 6.705286979675293, + "step": 180 + }, + { + "epoch": 2.5170305676855893, + "grad_norm": 0.011648476123809814, + "learning_rate": 0.002996173044030159, + "loss": 6.690052509307861, + "step": 181 + }, + { + "epoch": 2.531004366812227, + "grad_norm": 0.03942161798477173, + "learning_rate": 0.0029960768145808965, + "loss": 6.711885452270508, + "step": 182 + }, + { + "epoch": 2.544978165938865, + "grad_norm": 0.01571274921298027, + "learning_rate": 0.002995979391854797, + "loss": 6.715847969055176, + "step": 183 + }, + { + "epoch": 2.558951965065502, + "grad_norm": 0.00829714722931385, + "learning_rate": 0.0029958807759303515, + "loss": 6.685308933258057, + "step": 184 + }, + { + "epoch": 2.5729257641921395, + "grad_norm": 0.01374727487564087, + "learning_rate": 0.0029957809668870114, + "loss": 6.695744514465332, + "step": 185 + }, + { + "epoch": 2.5868995633187772, + "grad_norm": 0.010857059620320797, + "learning_rate": 0.0029956799648051934, + "loss": 6.658294677734375, + "step": 186 + }, + { + "epoch": 2.600873362445415, + "grad_norm": 0.008897246792912483, + "learning_rate": 0.002995577769766272, + "loss": 6.642439842224121, + "step": 187 + }, + { + "epoch": 2.6148471615720523, + "grad_norm": 0.015880772843956947, + "learning_rate": 0.002995474381852584, + "loss": 6.674907684326172, + "step": 188 + }, + { + "epoch": 2.62882096069869, + "grad_norm": 0.021623898297548294, + "learning_rate": 0.0029953698011474265, + "loss": 6.66236686706543, + "step": 189 + }, + { + "epoch": 2.6427947598253274, + "grad_norm": 0.03084627166390419, + "learning_rate": 0.0029952640277350596, + "loss": 6.666715621948242, + "step": 190 + }, + { + "epoch": 2.656768558951965, + "grad_norm": 0.040561504662036896, + "learning_rate": 0.002995157061700702, + "loss": 6.649062156677246, + "step": 191 + }, + { + "epoch": 2.6707423580786025, + "grad_norm": 0.031126515939831734, + "learning_rate": 0.0029950489031305345, + "loss": 6.640542030334473, + "step": 192 + }, + { + "epoch": 2.6847161572052403, + "grad_norm": 0.00872360821813345, + "learning_rate": 0.0029949395521116987, + "loss": 6.635221481323242, + "step": 193 + }, + { + "epoch": 2.6986899563318776, + "grad_norm": 0.023432321846485138, + "learning_rate": 0.0029948290087322966, + "loss": 6.642196178436279, + "step": 194 + }, + { + "epoch": 2.7126637554585153, + "grad_norm": 0.021096013486385345, + "learning_rate": 0.0029947172730813913, + "loss": 6.643655300140381, + "step": 195 + }, + { + "epoch": 2.726637554585153, + "grad_norm": 0.014494858682155609, + "learning_rate": 0.002994604345249006, + "loss": 6.612720966339111, + "step": 196 + }, + { + "epoch": 2.7406113537117904, + "grad_norm": 0.05225599184632301, + "learning_rate": 0.0029944902253261248, + "loss": 6.60659646987915, + "step": 197 + }, + { + "epoch": 2.7545851528384278, + "grad_norm": 0.014311269856989384, + "learning_rate": 0.0029943749134046915, + "loss": 6.54373025894165, + "step": 198 + }, + { + "epoch": 2.7685589519650655, + "grad_norm": 0.010777472518384457, + "learning_rate": 0.0029942584095776115, + "loss": 6.599837303161621, + "step": 199 + }, + { + "epoch": 2.7825327510917033, + "grad_norm": 0.012498079799115658, + "learning_rate": 0.002994140713938749, + "loss": 6.589657783508301, + "step": 200 + }, + { + "epoch": 2.7965065502183406, + "grad_norm": 0.015775132924318314, + "learning_rate": 0.00299402182658293, + "loss": 6.577511787414551, + "step": 201 + }, + { + "epoch": 2.810480349344978, + "grad_norm": 0.017978230491280556, + "learning_rate": 0.0029939017476059396, + "loss": 6.58717155456543, + "step": 202 + }, + { + "epoch": 2.8244541484716157, + "grad_norm": 0.026807745918631554, + "learning_rate": 0.0029937804771045236, + "loss": 6.578145503997803, + "step": 203 + }, + { + "epoch": 2.8384279475982535, + "grad_norm": 0.04853282868862152, + "learning_rate": 0.0029936580151763872, + "loss": 6.609755516052246, + "step": 204 + }, + { + "epoch": 2.8524017467248908, + "grad_norm": 0.1057916209101677, + "learning_rate": 0.002993534361920195, + "loss": 6.870367050170898, + "step": 205 + }, + { + "epoch": 2.8663755458515285, + "grad_norm": 0.04925911873579025, + "learning_rate": 0.002993409517435574, + "loss": 6.871556282043457, + "step": 206 + }, + { + "epoch": 2.880349344978166, + "grad_norm": 0.05288556218147278, + "learning_rate": 0.0029932834818231077, + "loss": 6.808315753936768, + "step": 207 + }, + { + "epoch": 2.8943231441048036, + "grad_norm": 0.03831710293889046, + "learning_rate": 0.0029931562551843405, + "loss": 6.827617645263672, + "step": 208 + }, + { + "epoch": 2.908296943231441, + "grad_norm": 0.03665117174386978, + "learning_rate": 0.002993027837621778, + "loss": 6.767284870147705, + "step": 209 + }, + { + "epoch": 2.9222707423580787, + "grad_norm": 0.026216736063361168, + "learning_rate": 0.0029928982292388826, + "loss": 6.733659744262695, + "step": 210 + }, + { + "epoch": 2.936244541484716, + "grad_norm": 0.024796854704618454, + "learning_rate": 0.002992767430140078, + "loss": 6.759577751159668, + "step": 211 + }, + { + "epoch": 2.950218340611354, + "grad_norm": 0.02410992607474327, + "learning_rate": 0.002992635440430747, + "loss": 6.701825141906738, + "step": 212 + }, + { + "epoch": 2.964192139737991, + "grad_norm": 0.019924134016036987, + "learning_rate": 0.0029925022602172308, + "loss": 6.677121162414551, + "step": 213 + }, + { + "epoch": 2.978165938864629, + "grad_norm": 0.022516867145895958, + "learning_rate": 0.00299236788960683, + "loss": 6.668622970581055, + "step": 214 + }, + { + "epoch": 2.992139737991266, + "grad_norm": 0.025068556889891624, + "learning_rate": 0.002992232328707806, + "loss": 6.67718505859375, + "step": 215 + }, + { + "epoch": 3.0, + "grad_norm": 0.02574208751320839, + "learning_rate": 0.002992095577629376, + "loss": 6.679594039916992, + "step": 216 + }, + { + "epoch": 3.0, + "eval_loss": 6.657555103302002, + "eval_runtime": 51.0199, + "eval_samples_per_second": 47.864, + "eval_steps_per_second": 1.509, + "step": 216 + }, + { + "epoch": 3.0139737991266378, + "grad_norm": 0.017298126593232155, + "learning_rate": 0.0029919576364817184, + "loss": 6.620774269104004, + "step": 217 + }, + { + "epoch": 3.027947598253275, + "grad_norm": 0.018840325996279716, + "learning_rate": 0.002991818505375971, + "loss": 6.613316535949707, + "step": 218 + }, + { + "epoch": 3.041921397379913, + "grad_norm": 0.020813113078475, + "learning_rate": 0.002991678184424228, + "loss": 6.621555328369141, + "step": 219 + }, + { + "epoch": 3.05589519650655, + "grad_norm": 0.026581479236483574, + "learning_rate": 0.0029915366737395437, + "loss": 6.613471031188965, + "step": 220 + }, + { + "epoch": 3.069868995633188, + "grad_norm": 0.02084030583500862, + "learning_rate": 0.0029913939734359306, + "loss": 6.559704780578613, + "step": 221 + }, + { + "epoch": 3.0838427947598253, + "grad_norm": 0.01327502354979515, + "learning_rate": 0.00299125008362836, + "loss": 6.564910888671875, + "step": 222 + }, + { + "epoch": 3.097816593886463, + "grad_norm": 0.015091734007000923, + "learning_rate": 0.0029911050044327614, + "loss": 6.571262836456299, + "step": 223 + }, + { + "epoch": 3.1117903930131003, + "grad_norm": 0.015463982708752155, + "learning_rate": 0.002990958735966023, + "loss": 6.565750598907471, + "step": 224 + }, + { + "epoch": 3.125764192139738, + "grad_norm": 0.013636826537549496, + "learning_rate": 0.0029908112783459894, + "loss": 6.57981014251709, + "step": 225 + }, + { + "epoch": 3.1397379912663754, + "grad_norm": 0.010613273829221725, + "learning_rate": 0.0029906626316914655, + "loss": 6.538505554199219, + "step": 226 + }, + { + "epoch": 3.153711790393013, + "grad_norm": 0.01804237812757492, + "learning_rate": 0.0029905127961222127, + "loss": 6.5282511711120605, + "step": 227 + }, + { + "epoch": 3.1676855895196505, + "grad_norm": 0.015131509862840176, + "learning_rate": 0.0029903617717589506, + "loss": 6.531335830688477, + "step": 228 + }, + { + "epoch": 3.1816593886462883, + "grad_norm": 0.010518206283450127, + "learning_rate": 0.002990209558723358, + "loss": 6.521886825561523, + "step": 229 + }, + { + "epoch": 3.1956331877729256, + "grad_norm": 0.011360022239387035, + "learning_rate": 0.002990056157138069, + "loss": 6.528301239013672, + "step": 230 + }, + { + "epoch": 3.2096069868995634, + "grad_norm": 0.011556181125342846, + "learning_rate": 0.0029899015671266773, + "loss": 6.497917652130127, + "step": 231 + }, + { + "epoch": 3.2235807860262007, + "grad_norm": 0.01365791168063879, + "learning_rate": 0.0029897457888137325, + "loss": 6.52849006652832, + "step": 232 + }, + { + "epoch": 3.2375545851528384, + "grad_norm": 0.01091043371707201, + "learning_rate": 0.0029895888223247426, + "loss": 6.492197036743164, + "step": 233 + }, + { + "epoch": 3.251528384279476, + "grad_norm": 0.008484847843647003, + "learning_rate": 0.002989430667786173, + "loss": 6.497377395629883, + "step": 234 + }, + { + "epoch": 3.2655021834061135, + "grad_norm": 0.00836879201233387, + "learning_rate": 0.0029892713253254466, + "loss": 6.495189666748047, + "step": 235 + }, + { + "epoch": 3.279475982532751, + "grad_norm": 0.009781497530639172, + "learning_rate": 0.0029891107950709406, + "loss": 6.454085826873779, + "step": 236 + }, + { + "epoch": 3.2934497816593886, + "grad_norm": 0.009603648446500301, + "learning_rate": 0.0029889490771519934, + "loss": 6.499475479125977, + "step": 237 + }, + { + "epoch": 3.3074235807860264, + "grad_norm": 0.007991855964064598, + "learning_rate": 0.0029887861716988975, + "loss": 6.460785865783691, + "step": 238 + }, + { + "epoch": 3.3213973799126637, + "grad_norm": 0.008447187952697277, + "learning_rate": 0.0029886220788429025, + "loss": 6.504582405090332, + "step": 239 + }, + { + "epoch": 3.3353711790393015, + "grad_norm": 0.010945498012006283, + "learning_rate": 0.0029884567987162156, + "loss": 6.49589729309082, + "step": 240 + }, + { + "epoch": 3.349344978165939, + "grad_norm": 0.011096215806901455, + "learning_rate": 0.0029882903314519997, + "loss": 6.460657119750977, + "step": 241 + }, + { + "epoch": 3.3633187772925766, + "grad_norm": 0.013094739057123661, + "learning_rate": 0.0029881226771843747, + "loss": 6.462450981140137, + "step": 242 + }, + { + "epoch": 3.377292576419214, + "grad_norm": 0.015423037111759186, + "learning_rate": 0.0029879538360484165, + "loss": 6.45565128326416, + "step": 243 + }, + { + "epoch": 3.3912663755458516, + "grad_norm": 0.0182222630828619, + "learning_rate": 0.002987783808180157, + "loss": 6.452227592468262, + "step": 244 + }, + { + "epoch": 3.405240174672489, + "grad_norm": 0.01981227844953537, + "learning_rate": 0.002987612593716585, + "loss": 6.477701663970947, + "step": 245 + }, + { + "epoch": 3.4192139737991267, + "grad_norm": 0.026733700186014175, + "learning_rate": 0.002987440192795645, + "loss": 6.452212333679199, + "step": 246 + }, + { + "epoch": 3.433187772925764, + "grad_norm": 0.03504155948758125, + "learning_rate": 0.0029872666055562372, + "loss": 6.509665489196777, + "step": 247 + }, + { + "epoch": 3.447161572052402, + "grad_norm": 0.027970125898718834, + "learning_rate": 0.002987091832138217, + "loss": 6.487089157104492, + "step": 248 + }, + { + "epoch": 3.461135371179039, + "grad_norm": 0.013714591972529888, + "learning_rate": 0.0029869158726823974, + "loss": 6.469569206237793, + "step": 249 + }, + { + "epoch": 3.475109170305677, + "grad_norm": 0.016472695395350456, + "learning_rate": 0.0029867387273305453, + "loss": 6.461688995361328, + "step": 250 + }, + { + "epoch": 3.489082969432314, + "grad_norm": 0.018457358703017235, + "learning_rate": 0.0029865603962253833, + "loss": 6.490985870361328, + "step": 251 + }, + { + "epoch": 3.503056768558952, + "grad_norm": 0.011045100167393684, + "learning_rate": 0.0029863808795105896, + "loss": 6.403701305389404, + "step": 252 + }, + { + "epoch": 3.5170305676855893, + "grad_norm": 0.014126897789537907, + "learning_rate": 0.0029862001773307972, + "loss": 6.379804611206055, + "step": 253 + }, + { + "epoch": 3.531004366812227, + "grad_norm": 0.014015335589647293, + "learning_rate": 0.0029860182898315954, + "loss": 6.404377460479736, + "step": 254 + }, + { + "epoch": 3.544978165938865, + "grad_norm": 0.010032658465206623, + "learning_rate": 0.002985835217159527, + "loss": 6.40291690826416, + "step": 255 + }, + { + "epoch": 3.558951965065502, + "grad_norm": 0.015469838865101337, + "learning_rate": 0.0029856509594620912, + "loss": 6.4414520263671875, + "step": 256 + }, + { + "epoch": 3.5729257641921395, + "grad_norm": 0.011106910184025764, + "learning_rate": 0.0029854655168877397, + "loss": 6.396444320678711, + "step": 257 + }, + { + "epoch": 3.5868995633187772, + "grad_norm": 0.01262320950627327, + "learning_rate": 0.0029852788895858816, + "loss": 6.385222434997559, + "step": 258 + }, + { + "epoch": 3.600873362445415, + "grad_norm": 0.012783976271748543, + "learning_rate": 0.002985091077706878, + "loss": 6.3818135261535645, + "step": 259 + }, + { + "epoch": 3.6148471615720523, + "grad_norm": 0.01224067248404026, + "learning_rate": 0.0029849020814020468, + "loss": 6.37886905670166, + "step": 260 + }, + { + "epoch": 3.62882096069869, + "grad_norm": 0.010540086776018143, + "learning_rate": 0.0029847119008236576, + "loss": 6.37799072265625, + "step": 261 + }, + { + "epoch": 3.6427947598253274, + "grad_norm": 0.012479913420975208, + "learning_rate": 0.0029845205361249366, + "loss": 6.365094184875488, + "step": 262 + }, + { + "epoch": 3.656768558951965, + "grad_norm": 0.010080077685415745, + "learning_rate": 0.0029843279874600622, + "loss": 6.371852874755859, + "step": 263 + }, + { + "epoch": 3.6707423580786025, + "grad_norm": 0.013266298919916153, + "learning_rate": 0.002984134254984168, + "loss": 6.380039215087891, + "step": 264 + }, + { + "epoch": 3.6847161572052403, + "grad_norm": 0.01014771405607462, + "learning_rate": 0.002983939338853339, + "loss": 6.370337963104248, + "step": 265 + }, + { + "epoch": 3.6986899563318776, + "grad_norm": 0.008164579048752785, + "learning_rate": 0.0029837432392246185, + "loss": 6.331899642944336, + "step": 266 + }, + { + "epoch": 3.7126637554585153, + "grad_norm": 0.01020442508161068, + "learning_rate": 0.002983545956255998, + "loss": 6.35077428817749, + "step": 267 + }, + { + "epoch": 3.726637554585153, + "grad_norm": 0.009898579679429531, + "learning_rate": 0.002983347490106426, + "loss": 6.371985912322998, + "step": 268 + }, + { + "epoch": 3.7406113537117904, + "grad_norm": 0.009668919257819653, + "learning_rate": 0.002983147840935802, + "loss": 6.356776237487793, + "step": 269 + }, + { + "epoch": 3.7545851528384278, + "grad_norm": 0.012560366652905941, + "learning_rate": 0.0029829470089049815, + "loss": 6.357911109924316, + "step": 270 + }, + { + "epoch": 3.7685589519650655, + "grad_norm": 0.018435249105095863, + "learning_rate": 0.0029827449941757698, + "loss": 6.366821765899658, + "step": 271 + }, + { + "epoch": 3.7825327510917033, + "grad_norm": 0.029046855866909027, + "learning_rate": 0.002982541796910927, + "loss": 6.367595672607422, + "step": 272 + }, + { + "epoch": 3.7965065502183406, + "grad_norm": 0.043010637164115906, + "learning_rate": 0.002982337417274165, + "loss": 6.413588523864746, + "step": 273 + }, + { + "epoch": 3.810480349344978, + "grad_norm": 0.045470818877220154, + "learning_rate": 0.0029821318554301497, + "loss": 6.435015678405762, + "step": 274 + }, + { + "epoch": 3.8244541484716157, + "grad_norm": 0.02148726023733616, + "learning_rate": 0.002981925111544498, + "loss": 6.330894470214844, + "step": 275 + }, + { + "epoch": 3.8384279475982535, + "grad_norm": 0.026808688417077065, + "learning_rate": 0.00298171718578378, + "loss": 6.384839057922363, + "step": 276 + }, + { + "epoch": 3.8524017467248908, + "grad_norm": 0.018495213240385056, + "learning_rate": 0.002981508078315517, + "loss": 6.36277961730957, + "step": 277 + }, + { + "epoch": 3.8663755458515285, + "grad_norm": 0.025393597781658173, + "learning_rate": 0.0029812977893081834, + "loss": 6.3702239990234375, + "step": 278 + }, + { + "epoch": 3.880349344978166, + "grad_norm": 0.020856289193034172, + "learning_rate": 0.002981086318931206, + "loss": 6.354547500610352, + "step": 279 + }, + { + "epoch": 3.8943231441048036, + "grad_norm": 0.017827816307544708, + "learning_rate": 0.002980873667354962, + "loss": 6.339264869689941, + "step": 280 + }, + { + "epoch": 3.908296943231441, + "grad_norm": 0.02639714442193508, + "learning_rate": 0.002980659834750781, + "loss": 6.3254194259643555, + "step": 281 + }, + { + "epoch": 3.9222707423580787, + "grad_norm": 0.015238096937537193, + "learning_rate": 0.002980444821290944, + "loss": 6.365842819213867, + "step": 282 + }, + { + "epoch": 3.936244541484716, + "grad_norm": 0.012863729149103165, + "learning_rate": 0.002980228627148684, + "loss": 6.356776714324951, + "step": 283 + }, + { + "epoch": 3.950218340611354, + "grad_norm": 0.014739899896085262, + "learning_rate": 0.0029800112524981845, + "loss": 6.332481384277344, + "step": 284 + }, + { + "epoch": 3.964192139737991, + "grad_norm": 0.013582171872258186, + "learning_rate": 0.0029797926975145797, + "loss": 6.311968803405762, + "step": 285 + }, + { + "epoch": 3.978165938864629, + "grad_norm": 0.012704637832939625, + "learning_rate": 0.0029795729623739555, + "loss": 6.321654319763184, + "step": 286 + }, + { + "epoch": 3.992139737991266, + "grad_norm": 0.012550560757517815, + "learning_rate": 0.0029793520472533495, + "loss": 6.298409461975098, + "step": 287 + }, + { + "epoch": 4.0, + "grad_norm": 0.011170165613293648, + "learning_rate": 0.002979129952330748, + "loss": 6.294830799102783, + "step": 288 + }, + { + "epoch": 4.0, + "eval_loss": 6.326757907867432, + "eval_runtime": 51.6566, + "eval_samples_per_second": 47.274, + "eval_steps_per_second": 1.491, + "step": 288 + }, + { + "epoch": 4.013973799126638, + "grad_norm": 0.010947446338832378, + "learning_rate": 0.0029789066777850897, + "loss": 6.286581039428711, + "step": 289 + }, + { + "epoch": 4.0279475982532755, + "grad_norm": 0.010097995400428772, + "learning_rate": 0.002978682223796262, + "loss": 6.288301467895508, + "step": 290 + }, + { + "epoch": 4.041921397379912, + "grad_norm": 0.01224475260823965, + "learning_rate": 0.002978456590545104, + "loss": 6.264986991882324, + "step": 291 + }, + { + "epoch": 4.05589519650655, + "grad_norm": 0.009734537452459335, + "learning_rate": 0.0029782297782134045, + "loss": 6.304243087768555, + "step": 292 + }, + { + "epoch": 4.069868995633188, + "grad_norm": 0.01136956550180912, + "learning_rate": 0.0029780017869839013, + "loss": 6.244938373565674, + "step": 293 + }, + { + "epoch": 4.083842794759826, + "grad_norm": 0.00935293547809124, + "learning_rate": 0.002977772617040284, + "loss": 6.251852035522461, + "step": 294 + }, + { + "epoch": 4.097816593886463, + "grad_norm": 0.011118784546852112, + "learning_rate": 0.0029775422685671894, + "loss": 6.256633758544922, + "step": 295 + }, + { + "epoch": 4.1117903930131, + "grad_norm": 0.010837923735380173, + "learning_rate": 0.0029773107417502055, + "loss": 6.263411045074463, + "step": 296 + }, + { + "epoch": 4.125764192139738, + "grad_norm": 0.011920305900275707, + "learning_rate": 0.0029770780367758703, + "loss": 6.239311218261719, + "step": 297 + }, + { + "epoch": 4.139737991266376, + "grad_norm": 0.014117483049631119, + "learning_rate": 0.0029768441538316685, + "loss": 6.253314018249512, + "step": 298 + }, + { + "epoch": 4.153711790393013, + "grad_norm": 0.022502148523926735, + "learning_rate": 0.002976609093106036, + "loss": 6.246156692504883, + "step": 299 + }, + { + "epoch": 4.1676855895196505, + "grad_norm": 0.03348720073699951, + "learning_rate": 0.002976372854788357, + "loss": 6.255914688110352, + "step": 300 + }, + { + "epoch": 4.181659388646288, + "grad_norm": 0.03427121415734291, + "learning_rate": 0.002976135439068965, + "loss": 6.289946556091309, + "step": 301 + }, + { + "epoch": 4.195633187772926, + "grad_norm": 0.02407355234026909, + "learning_rate": 0.002975896846139141, + "loss": 6.288534164428711, + "step": 302 + }, + { + "epoch": 4.209606986899563, + "grad_norm": 0.017861567437648773, + "learning_rate": 0.0029756570761911147, + "loss": 6.258049964904785, + "step": 303 + }, + { + "epoch": 4.223580786026201, + "grad_norm": 0.019980227574706078, + "learning_rate": 0.0029754161294180646, + "loss": 6.250895023345947, + "step": 304 + }, + { + "epoch": 4.2375545851528384, + "grad_norm": 0.01443127728998661, + "learning_rate": 0.0029751740060141177, + "loss": 6.212200164794922, + "step": 305 + }, + { + "epoch": 4.251528384279476, + "grad_norm": 0.014975055120885372, + "learning_rate": 0.0029749307061743475, + "loss": 6.230568885803223, + "step": 306 + }, + { + "epoch": 4.265502183406113, + "grad_norm": 0.01408296450972557, + "learning_rate": 0.0029746862300947773, + "loss": 6.243950366973877, + "step": 307 + }, + { + "epoch": 4.279475982532751, + "grad_norm": 0.011649327352643013, + "learning_rate": 0.0029744405779723766, + "loss": 6.197744846343994, + "step": 308 + }, + { + "epoch": 4.293449781659389, + "grad_norm": 0.013392188586294651, + "learning_rate": 0.002974193750005063, + "loss": 6.207277297973633, + "step": 309 + }, + { + "epoch": 4.307423580786026, + "grad_norm": 0.010902997106313705, + "learning_rate": 0.002973945746391701, + "loss": 6.166373252868652, + "step": 310 + }, + { + "epoch": 4.321397379912664, + "grad_norm": 0.012915392406284809, + "learning_rate": 0.0029736965673321032, + "loss": 6.186316013336182, + "step": 311 + }, + { + "epoch": 4.335371179039301, + "grad_norm": 0.010772830806672573, + "learning_rate": 0.0029734462130270273, + "loss": 6.172578811645508, + "step": 312 + }, + { + "epoch": 4.349344978165939, + "grad_norm": 0.010274911299347878, + "learning_rate": 0.002973194683678181, + "loss": 6.206851005554199, + "step": 313 + }, + { + "epoch": 4.3633187772925766, + "grad_norm": 0.013210494071245193, + "learning_rate": 0.0029729419794882164, + "loss": 6.21115779876709, + "step": 314 + }, + { + "epoch": 4.377292576419214, + "grad_norm": 0.013372313231229782, + "learning_rate": 0.0029726881006607326, + "loss": 6.18442440032959, + "step": 315 + }, + { + "epoch": 4.391266375545851, + "grad_norm": 0.008105908520519733, + "learning_rate": 0.0029724330474002743, + "loss": 6.204354286193848, + "step": 316 + }, + { + "epoch": 4.405240174672489, + "grad_norm": 0.014167108573019505, + "learning_rate": 0.0029721768199123346, + "loss": 6.180662155151367, + "step": 317 + }, + { + "epoch": 4.419213973799127, + "grad_norm": 0.008339803665876389, + "learning_rate": 0.002971919418403351, + "loss": 6.150079727172852, + "step": 318 + }, + { + "epoch": 4.4331877729257645, + "grad_norm": 0.015939386561512947, + "learning_rate": 0.0029716608430807073, + "loss": 6.175203323364258, + "step": 319 + }, + { + "epoch": 4.447161572052401, + "grad_norm": 0.010777109302580357, + "learning_rate": 0.0029714010941527327, + "loss": 6.166106224060059, + "step": 320 + }, + { + "epoch": 4.461135371179039, + "grad_norm": 0.011274735443294048, + "learning_rate": 0.002971140171828702, + "loss": 6.163111686706543, + "step": 321 + }, + { + "epoch": 4.475109170305677, + "grad_norm": 0.013626808300614357, + "learning_rate": 0.002970878076318836, + "loss": 6.140850067138672, + "step": 322 + }, + { + "epoch": 4.489082969432315, + "grad_norm": 0.016907762736082077, + "learning_rate": 0.0029706148078343, + "loss": 6.192012786865234, + "step": 323 + }, + { + "epoch": 4.503056768558952, + "grad_norm": 0.018817441537976265, + "learning_rate": 0.002970350366587205, + "loss": 6.162238121032715, + "step": 324 + }, + { + "epoch": 4.517030567685589, + "grad_norm": 0.02829768694937229, + "learning_rate": 0.0029700847527906068, + "loss": 6.159066200256348, + "step": 325 + }, + { + "epoch": 4.531004366812227, + "grad_norm": 0.02034052088856697, + "learning_rate": 0.002969817966658505, + "loss": 6.164158344268799, + "step": 326 + }, + { + "epoch": 4.544978165938865, + "grad_norm": 0.01816369965672493, + "learning_rate": 0.0029695500084058447, + "loss": 6.132316589355469, + "step": 327 + }, + { + "epoch": 4.558951965065502, + "grad_norm": 0.013685261830687523, + "learning_rate": 0.0029692808782485146, + "loss": 6.152181625366211, + "step": 328 + }, + { + "epoch": 4.5729257641921395, + "grad_norm": 0.009791149757802486, + "learning_rate": 0.0029690105764033485, + "loss": 6.138544082641602, + "step": 329 + }, + { + "epoch": 4.586899563318777, + "grad_norm": 0.014827069826424122, + "learning_rate": 0.002968739103088124, + "loss": 6.135766983032227, + "step": 330 + }, + { + "epoch": 4.600873362445415, + "grad_norm": 0.01391249243170023, + "learning_rate": 0.0029684664585215618, + "loss": 6.146172523498535, + "step": 331 + }, + { + "epoch": 4.614847161572053, + "grad_norm": 0.012729899026453495, + "learning_rate": 0.002968192642923326, + "loss": 6.122613430023193, + "step": 332 + }, + { + "epoch": 4.62882096069869, + "grad_norm": 0.014175458811223507, + "learning_rate": 0.002967917656514026, + "loss": 6.12574577331543, + "step": 333 + }, + { + "epoch": 4.642794759825327, + "grad_norm": 0.010958840139210224, + "learning_rate": 0.002967641499515213, + "loss": 6.111088752746582, + "step": 334 + }, + { + "epoch": 4.656768558951965, + "grad_norm": 0.015380816534161568, + "learning_rate": 0.0029673641721493817, + "loss": 6.100615501403809, + "step": 335 + }, + { + "epoch": 4.670742358078603, + "grad_norm": 0.025339096784591675, + "learning_rate": 0.00296708567463997, + "loss": 6.139659881591797, + "step": 336 + }, + { + "epoch": 4.68471615720524, + "grad_norm": 0.03695995733141899, + "learning_rate": 0.002966806007211358, + "loss": 6.160996437072754, + "step": 337 + }, + { + "epoch": 4.698689956331878, + "grad_norm": 0.0408046580851078, + "learning_rate": 0.0029665251700888686, + "loss": 6.173980712890625, + "step": 338 + }, + { + "epoch": 4.712663755458515, + "grad_norm": 0.025022024288773537, + "learning_rate": 0.0029662431634987674, + "loss": 6.145794868469238, + "step": 339 + }, + { + "epoch": 4.726637554585153, + "grad_norm": 0.027254154905676842, + "learning_rate": 0.002965959987668262, + "loss": 6.16928243637085, + "step": 340 + }, + { + "epoch": 4.74061135371179, + "grad_norm": 0.06191903352737427, + "learning_rate": 0.0029656756428255017, + "loss": 6.138525009155273, + "step": 341 + }, + { + "epoch": 4.754585152838428, + "grad_norm": 0.0348956398665905, + "learning_rate": 0.002965390129199579, + "loss": 6.185477256774902, + "step": 342 + }, + { + "epoch": 4.7685589519650655, + "grad_norm": 0.020859677344560623, + "learning_rate": 0.002965103447020526, + "loss": 6.148846626281738, + "step": 343 + }, + { + "epoch": 4.782532751091703, + "grad_norm": 0.022810814902186394, + "learning_rate": 0.0029648155965193174, + "loss": 6.1701178550720215, + "step": 344 + }, + { + "epoch": 4.796506550218341, + "grad_norm": 0.017981817945837975, + "learning_rate": 0.00296452657792787, + "loss": 6.1323771476745605, + "step": 345 + }, + { + "epoch": 4.810480349344978, + "grad_norm": 0.014905866235494614, + "learning_rate": 0.00296423639147904, + "loss": 6.123426914215088, + "step": 346 + }, + { + "epoch": 4.824454148471616, + "grad_norm": 0.01463468000292778, + "learning_rate": 0.002963945037406626, + "loss": 6.146122932434082, + "step": 347 + }, + { + "epoch": 4.8384279475982535, + "grad_norm": 0.021616583690047264, + "learning_rate": 0.002963652515945366, + "loss": 6.093459606170654, + "step": 348 + }, + { + "epoch": 4.85240174672489, + "grad_norm": 0.01550968922674656, + "learning_rate": 0.00296335882733094, + "loss": 6.114929676055908, + "step": 349 + }, + { + "epoch": 4.866375545851528, + "grad_norm": 0.014597428031265736, + "learning_rate": 0.0029630639717999675, + "loss": 6.113945960998535, + "step": 350 + }, + { + "epoch": 4.880349344978166, + "grad_norm": 0.018249280750751495, + "learning_rate": 0.002962767949590008, + "loss": 6.130507469177246, + "step": 351 + }, + { + "epoch": 4.894323144104804, + "grad_norm": 0.014160148799419403, + "learning_rate": 0.0029624707609395614, + "loss": 6.086946487426758, + "step": 352 + }, + { + "epoch": 4.908296943231441, + "grad_norm": 0.015563832595944405, + "learning_rate": 0.002962172406088067, + "loss": 6.083332538604736, + "step": 353 + }, + { + "epoch": 4.922270742358078, + "grad_norm": 0.012240015901625156, + "learning_rate": 0.002961872885275904, + "loss": 6.0822906494140625, + "step": 354 + }, + { + "epoch": 4.936244541484716, + "grad_norm": 0.0141651201993227, + "learning_rate": 0.002961572198744391, + "loss": 6.049562454223633, + "step": 355 + }, + { + "epoch": 4.950218340611354, + "grad_norm": 0.010004645213484764, + "learning_rate": 0.0029612703467357866, + "loss": 6.067063808441162, + "step": 356 + }, + { + "epoch": 4.964192139737992, + "grad_norm": 0.011790835298597813, + "learning_rate": 0.0029609673294932856, + "loss": 6.062193870544434, + "step": 357 + }, + { + "epoch": 4.978165938864628, + "grad_norm": 0.011088060215115547, + "learning_rate": 0.002960663147261025, + "loss": 6.070950031280518, + "step": 358 + }, + { + "epoch": 4.992139737991266, + "grad_norm": 0.010387620888650417, + "learning_rate": 0.002960357800284078, + "loss": 6.032423973083496, + "step": 359 + }, + { + "epoch": 5.0, + "grad_norm": 0.012074458412826061, + "learning_rate": 0.002960051288808458, + "loss": 6.066788673400879, + "step": 360 + }, + { + "epoch": 5.0, + "eval_loss": 6.097111701965332, + "eval_runtime": 51.8742, + "eval_samples_per_second": 47.075, + "eval_steps_per_second": 1.484, + "step": 360 + }, + { + "epoch": 5.013973799126638, + "grad_norm": 0.013429952785372734, + "learning_rate": 0.0029597436130811153, + "loss": 6.022172451019287, + "step": 361 + }, + { + "epoch": 5.0279475982532755, + "grad_norm": 0.01764705777168274, + "learning_rate": 0.0029594347733499387, + "loss": 6.029745101928711, + "step": 362 + }, + { + "epoch": 5.041921397379912, + "grad_norm": 0.02403099089860916, + "learning_rate": 0.002959124769863755, + "loss": 6.04707145690918, + "step": 363 + }, + { + "epoch": 5.05589519650655, + "grad_norm": 0.029030559584498405, + "learning_rate": 0.002958813602872328, + "loss": 6.05479097366333, + "step": 364 + }, + { + "epoch": 5.069868995633188, + "grad_norm": 0.029980644583702087, + "learning_rate": 0.0029585012726263595, + "loss": 6.068657875061035, + "step": 365 + }, + { + "epoch": 5.083842794759826, + "grad_norm": 0.021424276754260063, + "learning_rate": 0.0029581877793774885, + "loss": 6.062008857727051, + "step": 366 + }, + { + "epoch": 5.097816593886463, + "grad_norm": 0.018076395615935326, + "learning_rate": 0.002957873123378291, + "loss": 6.023279190063477, + "step": 367 + }, + { + "epoch": 5.1117903930131, + "grad_norm": 0.017688792198896408, + "learning_rate": 0.002957557304882279, + "loss": 6.014641284942627, + "step": 368 + }, + { + "epoch": 5.125764192139738, + "grad_norm": 0.02313626930117607, + "learning_rate": 0.0029572403241439026, + "loss": 6.026420593261719, + "step": 369 + }, + { + "epoch": 5.139737991266376, + "grad_norm": 0.04045062139630318, + "learning_rate": 0.0029569221814185467, + "loss": 6.007199764251709, + "step": 370 + }, + { + "epoch": 5.153711790393013, + "grad_norm": 0.015792209655046463, + "learning_rate": 0.0029566028769625334, + "loss": 6.017209053039551, + "step": 371 + }, + { + "epoch": 5.1676855895196505, + "grad_norm": 0.01990479789674282, + "learning_rate": 0.002956282411033121, + "loss": 5.982841491699219, + "step": 372 + }, + { + "epoch": 5.181659388646288, + "grad_norm": 0.01191043108701706, + "learning_rate": 0.002955960783888503, + "loss": 6.0035929679870605, + "step": 373 + }, + { + "epoch": 5.195633187772926, + "grad_norm": 0.01732831820845604, + "learning_rate": 0.002955637995787807, + "loss": 6.025006294250488, + "step": 374 + }, + { + "epoch": 5.209606986899563, + "grad_norm": 0.032948266714811325, + "learning_rate": 0.0029553140469910995, + "loss": 6.007736682891846, + "step": 375 + }, + { + "epoch": 5.223580786026201, + "grad_norm": 0.011192727833986282, + "learning_rate": 0.00295498893775938, + "loss": 6.0022687911987305, + "step": 376 + }, + { + "epoch": 5.2375545851528384, + "grad_norm": 0.014646840281784534, + "learning_rate": 0.0029546626683545824, + "loss": 5.976479530334473, + "step": 377 + }, + { + "epoch": 5.251528384279476, + "grad_norm": 0.01589796133339405, + "learning_rate": 0.002954335239039576, + "loss": 5.980949878692627, + "step": 378 + }, + { + "epoch": 5.265502183406113, + "grad_norm": 0.015129734762012959, + "learning_rate": 0.0029540066500781657, + "loss": 6.0073394775390625, + "step": 379 + }, + { + "epoch": 5.279475982532751, + "grad_norm": 0.011140518821775913, + "learning_rate": 0.002953676901735088, + "loss": 5.976587295532227, + "step": 380 + }, + { + "epoch": 5.293449781659389, + "grad_norm": 0.0109480582177639, + "learning_rate": 0.0029533459942760174, + "loss": 6.001427173614502, + "step": 381 + }, + { + "epoch": 5.307423580786026, + "grad_norm": 0.013346790336072445, + "learning_rate": 0.0029530139279675583, + "loss": 5.96919059753418, + "step": 382 + }, + { + "epoch": 5.321397379912664, + "grad_norm": 0.012474581599235535, + "learning_rate": 0.002952680703077251, + "loss": 5.970426559448242, + "step": 383 + }, + { + "epoch": 5.335371179039301, + "grad_norm": 0.01140566635876894, + "learning_rate": 0.002952346319873569, + "loss": 5.989171504974365, + "step": 384 + }, + { + "epoch": 5.349344978165939, + "grad_norm": 0.0137029392644763, + "learning_rate": 0.0029520107786259187, + "loss": 6.010115623474121, + "step": 385 + }, + { + "epoch": 5.3633187772925766, + "grad_norm": 0.015517761930823326, + "learning_rate": 0.002951674079604639, + "loss": 6.003649711608887, + "step": 386 + }, + { + "epoch": 5.377292576419214, + "grad_norm": 0.02173629403114319, + "learning_rate": 0.0029513362230810035, + "loss": 5.967074871063232, + "step": 387 + }, + { + "epoch": 5.391266375545851, + "grad_norm": 0.02683914266526699, + "learning_rate": 0.002950997209327216, + "loss": 5.994434833526611, + "step": 388 + }, + { + "epoch": 5.405240174672489, + "grad_norm": 0.025865858420729637, + "learning_rate": 0.002950657038616415, + "loss": 6.010523796081543, + "step": 389 + }, + { + "epoch": 5.419213973799127, + "grad_norm": 0.02544897235929966, + "learning_rate": 0.002950315711222668, + "loss": 5.994078159332275, + "step": 390 + }, + { + "epoch": 5.4331877729257645, + "grad_norm": 0.028426019474864006, + "learning_rate": 0.0029499732274209777, + "loss": 6.020925521850586, + "step": 391 + }, + { + "epoch": 5.447161572052401, + "grad_norm": 0.026174750179052353, + "learning_rate": 0.0029496295874872768, + "loss": 5.985616683959961, + "step": 392 + }, + { + "epoch": 5.461135371179039, + "grad_norm": 0.02802138403058052, + "learning_rate": 0.00294928479169843, + "loss": 5.973659038543701, + "step": 393 + }, + { + "epoch": 5.475109170305677, + "grad_norm": 0.019348938018083572, + "learning_rate": 0.002948938840332233, + "loss": 5.990137577056885, + "step": 394 + }, + { + "epoch": 5.489082969432315, + "grad_norm": 0.016294041648507118, + "learning_rate": 0.0029485917336674115, + "loss": 5.962082862854004, + "step": 395 + }, + { + "epoch": 5.503056768558952, + "grad_norm": 0.016167785972356796, + "learning_rate": 0.002948243471983625, + "loss": 5.9594573974609375, + "step": 396 + }, + { + "epoch": 5.517030567685589, + "grad_norm": 0.014119232073426247, + "learning_rate": 0.00294789405556146, + "loss": 5.894467353820801, + "step": 397 + }, + { + "epoch": 5.531004366812227, + "grad_norm": 0.013805347494781017, + "learning_rate": 0.002947543484682436, + "loss": 5.964014530181885, + "step": 398 + }, + { + "epoch": 5.544978165938865, + "grad_norm": 0.014496497809886932, + "learning_rate": 0.002947191759629001, + "loss": 5.989920616149902, + "step": 399 + }, + { + "epoch": 5.558951965065502, + "grad_norm": 0.014813800342381, + "learning_rate": 0.0029468388806845333, + "loss": 5.9669413566589355, + "step": 400 + }, + { + "epoch": 5.5729257641921395, + "grad_norm": 0.01153622381389141, + "learning_rate": 0.0029464848481333423, + "loss": 5.918018341064453, + "step": 401 + }, + { + "epoch": 5.586899563318777, + "grad_norm": 0.010525977239012718, + "learning_rate": 0.0029461296622606647, + "loss": 5.892477989196777, + "step": 402 + }, + { + "epoch": 5.600873362445415, + "grad_norm": 0.011410463601350784, + "learning_rate": 0.002945773323352667, + "loss": 5.925249099731445, + "step": 403 + }, + { + "epoch": 5.614847161572053, + "grad_norm": 0.010949424467980862, + "learning_rate": 0.0029454158316964464, + "loss": 5.90645694732666, + "step": 404 + }, + { + "epoch": 5.62882096069869, + "grad_norm": 0.00962804164737463, + "learning_rate": 0.002945057187580026, + "loss": 5.917054176330566, + "step": 405 + }, + { + "epoch": 5.642794759825327, + "grad_norm": 0.011517546139657497, + "learning_rate": 0.0029446973912923597, + "loss": 5.919445514678955, + "step": 406 + }, + { + "epoch": 5.656768558951965, + "grad_norm": 0.01567666605114937, + "learning_rate": 0.0029443364431233285, + "loss": 5.884111404418945, + "step": 407 + }, + { + "epoch": 5.670742358078603, + "grad_norm": 0.016781140118837357, + "learning_rate": 0.002943974343363742, + "loss": 5.9252119064331055, + "step": 408 + }, + { + "epoch": 5.68471615720524, + "grad_norm": 0.02244546078145504, + "learning_rate": 0.0029436110923053375, + "loss": 5.933143615722656, + "step": 409 + }, + { + "epoch": 5.698689956331878, + "grad_norm": 0.028186792507767677, + "learning_rate": 0.00294324669024078, + "loss": 5.9079132080078125, + "step": 410 + }, + { + "epoch": 5.712663755458515, + "grad_norm": 0.028585223481059074, + "learning_rate": 0.002942881137463661, + "loss": 5.925262451171875, + "step": 411 + }, + { + "epoch": 5.726637554585153, + "grad_norm": 0.02781485952436924, + "learning_rate": 0.0029425144342685003, + "loss": 5.963315010070801, + "step": 412 + }, + { + "epoch": 5.74061135371179, + "grad_norm": 0.021486928686499596, + "learning_rate": 0.002942146580950744, + "loss": 5.930115699768066, + "step": 413 + }, + { + "epoch": 5.754585152838428, + "grad_norm": 0.016890818253159523, + "learning_rate": 0.002941777577806765, + "loss": 5.921439170837402, + "step": 414 + }, + { + "epoch": 5.7685589519650655, + "grad_norm": 0.0169126745313406, + "learning_rate": 0.0029414074251338618, + "loss": 5.928220272064209, + "step": 415 + }, + { + "epoch": 5.782532751091703, + "grad_norm": 0.014659176580607891, + "learning_rate": 0.002941036123230261, + "loss": 5.934546947479248, + "step": 416 + }, + { + "epoch": 5.796506550218341, + "grad_norm": 0.015337883494794369, + "learning_rate": 0.002940663672395112, + "loss": 5.890145301818848, + "step": 417 + }, + { + "epoch": 5.810480349344978, + "grad_norm": 0.01254988368600607, + "learning_rate": 0.002940290072928493, + "loss": 5.875492095947266, + "step": 418 + }, + { + "epoch": 5.824454148471616, + "grad_norm": 0.014658362604677677, + "learning_rate": 0.0029399153251314066, + "loss": 5.901272773742676, + "step": 419 + }, + { + "epoch": 5.8384279475982535, + "grad_norm": 0.017342250794172287, + "learning_rate": 0.002939539429305779, + "loss": 5.875853538513184, + "step": 420 + }, + { + "epoch": 5.85240174672489, + "grad_norm": 0.016825608909130096, + "learning_rate": 0.0029391623857544643, + "loss": 5.941972732543945, + "step": 421 + }, + { + "epoch": 5.866375545851528, + "grad_norm": 0.016716884449124336, + "learning_rate": 0.002938784194781238, + "loss": 5.912367343902588, + "step": 422 + }, + { + "epoch": 5.880349344978166, + "grad_norm": 0.021707067266106606, + "learning_rate": 0.002938404856690803, + "loss": 5.895319938659668, + "step": 423 + }, + { + "epoch": 5.894323144104804, + "grad_norm": 0.030519135296344757, + "learning_rate": 0.002938024371788785, + "loss": 5.897165298461914, + "step": 424 + }, + { + "epoch": 5.908296943231441, + "grad_norm": 0.02572188340127468, + "learning_rate": 0.002937642740381733, + "loss": 5.879164695739746, + "step": 425 + }, + { + "epoch": 5.922270742358078, + "grad_norm": 0.016286717727780342, + "learning_rate": 0.002937259962777122, + "loss": 5.841421127319336, + "step": 426 + }, + { + "epoch": 5.936244541484716, + "grad_norm": 0.020656082779169083, + "learning_rate": 0.002936876039283347, + "loss": 5.9222259521484375, + "step": 427 + }, + { + "epoch": 5.950218340611354, + "grad_norm": 0.015350459143519402, + "learning_rate": 0.0029364909702097298, + "loss": 5.885847091674805, + "step": 428 + }, + { + "epoch": 5.964192139737992, + "grad_norm": 0.01708570495247841, + "learning_rate": 0.0029361047558665136, + "loss": 5.872044563293457, + "step": 429 + }, + { + "epoch": 5.978165938864628, + "grad_norm": 0.013537376187741756, + "learning_rate": 0.0029357173965648634, + "loss": 5.866434097290039, + "step": 430 + }, + { + "epoch": 5.992139737991266, + "grad_norm": 0.013208637945353985, + "learning_rate": 0.002935328892616868, + "loss": 5.8624491691589355, + "step": 431 + }, + { + "epoch": 6.0, + "grad_norm": 0.012249554507434368, + "learning_rate": 0.002934939244335538, + "loss": 5.833232879638672, + "step": 432 + }, + { + "epoch": 6.0, + "eval_loss": 5.90985107421875, + "eval_runtime": 51.0696, + "eval_samples_per_second": 47.817, + "eval_steps_per_second": 1.508, + "step": 432 + }, + { + "epoch": 6.013973799126638, + "grad_norm": 0.01382362935692072, + "learning_rate": 0.0029345484520348062, + "loss": 5.835695266723633, + "step": 433 + }, + { + "epoch": 6.0279475982532755, + "grad_norm": 0.012947693467140198, + "learning_rate": 0.0029341565160295262, + "loss": 5.826190948486328, + "step": 434 + }, + { + "epoch": 6.041921397379912, + "grad_norm": 0.013653872534632683, + "learning_rate": 0.002933763436635475, + "loss": 5.831789016723633, + "step": 435 + }, + { + "epoch": 6.05589519650655, + "grad_norm": 0.027596980333328247, + "learning_rate": 0.0029333692141693484, + "loss": 5.839963436126709, + "step": 436 + }, + { + "epoch": 6.069868995633188, + "grad_norm": 0.01637990213930607, + "learning_rate": 0.0029329738489487648, + "loss": 5.805248260498047, + "step": 437 + }, + { + "epoch": 6.083842794759826, + "grad_norm": 0.02169182151556015, + "learning_rate": 0.0029325773412922627, + "loss": 5.838248252868652, + "step": 438 + }, + { + "epoch": 6.097816593886463, + "grad_norm": 0.03518201410770416, + "learning_rate": 0.0029321796915193015, + "loss": 5.834336280822754, + "step": 439 + }, + { + "epoch": 6.1117903930131, + "grad_norm": 0.04983877018094063, + "learning_rate": 0.0029317808999502603, + "loss": 5.879856109619141, + "step": 440 + }, + { + "epoch": 6.125764192139738, + "grad_norm": 0.05085664615035057, + "learning_rate": 0.0029313809669064377, + "loss": 5.946531295776367, + "step": 441 + }, + { + "epoch": 6.139737991266376, + "grad_norm": 0.03314858302474022, + "learning_rate": 0.0029309798927100532, + "loss": 5.909746170043945, + "step": 442 + }, + { + "epoch": 6.153711790393013, + "grad_norm": 0.034118980169296265, + "learning_rate": 0.0029305776776842453, + "loss": 5.903947830200195, + "step": 443 + }, + { + "epoch": 6.1676855895196505, + "grad_norm": 0.03401247411966324, + "learning_rate": 0.0029301743221530705, + "loss": 5.936162948608398, + "step": 444 + }, + { + "epoch": 6.181659388646288, + "grad_norm": 0.043827857822179794, + "learning_rate": 0.0029297698264415047, + "loss": 5.953949451446533, + "step": 445 + }, + { + "epoch": 6.195633187772926, + "grad_norm": 0.06281330436468124, + "learning_rate": 0.002929364190875444, + "loss": 5.988225936889648, + "step": 446 + }, + { + "epoch": 6.209606986899563, + "grad_norm": 0.07216024398803711, + "learning_rate": 0.0029289574157817014, + "loss": 6.187741279602051, + "step": 447 + }, + { + "epoch": 6.223580786026201, + "grad_norm": 0.055205658078193665, + "learning_rate": 0.002928549501488008, + "loss": 6.177831649780273, + "step": 448 + }, + { + "epoch": 6.2375545851528384, + "grad_norm": 0.04834286496043205, + "learning_rate": 0.0029281404483230133, + "loss": 6.209604263305664, + "step": 449 + }, + { + "epoch": 6.251528384279476, + "grad_norm": 0.0447777584195137, + "learning_rate": 0.0029277302566162837, + "loss": 6.146507263183594, + "step": 450 + }, + { + "epoch": 6.265502183406113, + "grad_norm": 0.05729430541396141, + "learning_rate": 0.0029273189266983035, + "loss": 6.280686378479004, + "step": 451 + }, + { + "epoch": 6.279475982532751, + "grad_norm": 0.0410807766020298, + "learning_rate": 0.002926906458900473, + "loss": 6.246824264526367, + "step": 452 + }, + { + "epoch": 6.293449781659389, + "grad_norm": 0.044607777148485184, + "learning_rate": 0.0029264928535551104, + "loss": 6.19956111907959, + "step": 453 + }, + { + "epoch": 6.307423580786026, + "grad_norm": 0.05141043663024902, + "learning_rate": 0.0029260781109954515, + "loss": 6.118578910827637, + "step": 454 + }, + { + "epoch": 6.321397379912664, + "grad_norm": 0.029248764738440514, + "learning_rate": 0.0029256622315556448, + "loss": 6.155360221862793, + "step": 455 + }, + { + "epoch": 6.335371179039301, + "grad_norm": 0.028403814882040024, + "learning_rate": 0.0029252452155707582, + "loss": 6.097352027893066, + "step": 456 + }, + { + "epoch": 6.349344978165939, + "grad_norm": 0.025872666388750076, + "learning_rate": 0.002924827063376773, + "loss": 6.072485446929932, + "step": 457 + }, + { + "epoch": 6.3633187772925766, + "grad_norm": 0.018447555601596832, + "learning_rate": 0.0029244077753105874, + "loss": 6.077045440673828, + "step": 458 + }, + { + "epoch": 6.377292576419214, + "grad_norm": 0.023166058585047722, + "learning_rate": 0.0029239873517100147, + "loss": 6.040050983428955, + "step": 459 + }, + { + "epoch": 6.391266375545851, + "grad_norm": 0.019579261541366577, + "learning_rate": 0.002923565792913782, + "loss": 6.039424896240234, + "step": 460 + }, + { + "epoch": 6.405240174672489, + "grad_norm": 0.018369073048233986, + "learning_rate": 0.002923143099261531, + "loss": 6.021180152893066, + "step": 461 + }, + { + "epoch": 6.419213973799127, + "grad_norm": 0.014188461937010288, + "learning_rate": 0.0029227192710938195, + "loss": 5.9668049812316895, + "step": 462 + }, + { + "epoch": 6.4331877729257645, + "grad_norm": 0.019567398354411125, + "learning_rate": 0.0029222943087521182, + "loss": 5.989189147949219, + "step": 463 + }, + { + "epoch": 6.447161572052401, + "grad_norm": 0.014190621674060822, + "learning_rate": 0.0029218682125788102, + "loss": 5.942726135253906, + "step": 464 + }, + { + "epoch": 6.461135371179039, + "grad_norm": 0.01375526562333107, + "learning_rate": 0.002921440982917195, + "loss": 5.942777633666992, + "step": 465 + }, + { + "epoch": 6.475109170305677, + "grad_norm": 0.014203300699591637, + "learning_rate": 0.0029210126201114836, + "loss": 5.927928924560547, + "step": 466 + }, + { + "epoch": 6.489082969432315, + "grad_norm": 0.014231069944798946, + "learning_rate": 0.0029205831245067994, + "loss": 5.923510551452637, + "step": 467 + }, + { + "epoch": 6.503056768558952, + "grad_norm": 0.012771515175700188, + "learning_rate": 0.0029201524964491796, + "loss": 5.908803939819336, + "step": 468 + }, + { + "epoch": 6.517030567685589, + "grad_norm": 0.013399476185441017, + "learning_rate": 0.0029197207362855735, + "loss": 5.890079021453857, + "step": 469 + }, + { + "epoch": 6.531004366812227, + "grad_norm": 0.025862423703074455, + "learning_rate": 0.0029192878443638427, + "loss": 5.919907569885254, + "step": 470 + }, + { + "epoch": 6.544978165938865, + "grad_norm": 0.011315986514091492, + "learning_rate": 0.0029188538210327595, + "loss": 5.901529788970947, + "step": 471 + }, + { + "epoch": 6.558951965065502, + "grad_norm": 0.01235947571694851, + "learning_rate": 0.00291841866664201, + "loss": 5.903595924377441, + "step": 472 + }, + { + "epoch": 6.5729257641921395, + "grad_norm": 0.012175127863883972, + "learning_rate": 0.0029179823815421893, + "loss": 5.868306636810303, + "step": 473 + }, + { + "epoch": 6.586899563318777, + "grad_norm": 0.01195445004850626, + "learning_rate": 0.0029175449660848052, + "loss": 5.899898529052734, + "step": 474 + }, + { + "epoch": 6.600873362445415, + "grad_norm": 0.012801481410861015, + "learning_rate": 0.0029171064206222745, + "loss": 5.890621662139893, + "step": 475 + }, + { + "epoch": 6.614847161572053, + "grad_norm": 0.011909718625247478, + "learning_rate": 0.002916666745507926, + "loss": 5.86421012878418, + "step": 476 + }, + { + "epoch": 6.62882096069869, + "grad_norm": 0.01162486057728529, + "learning_rate": 0.0029162259410959977, + "loss": 5.869139671325684, + "step": 477 + }, + { + "epoch": 6.642794759825327, + "grad_norm": 0.010923407040536404, + "learning_rate": 0.0029157840077416374, + "loss": 5.869789123535156, + "step": 478 + }, + { + "epoch": 6.656768558951965, + "grad_norm": 0.011943561024963856, + "learning_rate": 0.002915340945800904, + "loss": 5.873437881469727, + "step": 479 + }, + { + "epoch": 6.670742358078603, + "grad_norm": 0.010295520536601543, + "learning_rate": 0.002914896755630763, + "loss": 5.819953918457031, + "step": 480 + }, + { + "epoch": 6.68471615720524, + "grad_norm": 0.0093373479321599, + "learning_rate": 0.002914451437589092, + "loss": 5.829481601715088, + "step": 481 + }, + { + "epoch": 6.698689956331878, + "grad_norm": 0.01182442344725132, + "learning_rate": 0.002914004992034674, + "loss": 5.855249404907227, + "step": 482 + }, + { + "epoch": 6.712663755458515, + "grad_norm": 0.00939213763922453, + "learning_rate": 0.0029135574193272034, + "loss": 5.83876371383667, + "step": 483 + }, + { + "epoch": 6.726637554585153, + "grad_norm": 0.011648556217551231, + "learning_rate": 0.0029131087198272813, + "loss": 5.824906826019287, + "step": 484 + }, + { + "epoch": 6.74061135371179, + "grad_norm": 0.011227169074118137, + "learning_rate": 0.002912658893896416, + "loss": 5.847869873046875, + "step": 485 + }, + { + "epoch": 6.754585152838428, + "grad_norm": 0.011084851808845997, + "learning_rate": 0.002912207941897025, + "loss": 5.839183807373047, + "step": 486 + }, + { + "epoch": 6.7685589519650655, + "grad_norm": 0.009923825040459633, + "learning_rate": 0.002911755864192432, + "loss": 5.818821907043457, + "step": 487 + }, + { + "epoch": 6.782532751091703, + "grad_norm": 0.011300318874418736, + "learning_rate": 0.002911302661146867, + "loss": 5.819517135620117, + "step": 488 + }, + { + "epoch": 6.796506550218341, + "grad_norm": 0.009915645234286785, + "learning_rate": 0.0029108483331254684, + "loss": 5.776939392089844, + "step": 489 + }, + { + "epoch": 6.810480349344978, + "grad_norm": 0.008965205401182175, + "learning_rate": 0.00291039288049428, + "loss": 5.824148178100586, + "step": 490 + }, + { + "epoch": 6.824454148471616, + "grad_norm": 0.011025927029550076, + "learning_rate": 0.0029099363036202515, + "loss": 5.832704544067383, + "step": 491 + }, + { + "epoch": 6.8384279475982535, + "grad_norm": 0.010143411345779896, + "learning_rate": 0.0029094786028712387, + "loss": 5.802353382110596, + "step": 492 + }, + { + "epoch": 6.85240174672489, + "grad_norm": 0.010021256282925606, + "learning_rate": 0.002909019778616003, + "loss": 5.802136421203613, + "step": 493 + }, + { + "epoch": 6.866375545851528, + "grad_norm": 0.011356706731021404, + "learning_rate": 0.002908559831224211, + "loss": 5.807962417602539, + "step": 494 + }, + { + "epoch": 6.880349344978166, + "grad_norm": 0.010525298304855824, + "learning_rate": 0.002908098761066433, + "loss": 5.787661552429199, + "step": 495 + }, + { + "epoch": 6.894323144104804, + "grad_norm": 0.011147079057991505, + "learning_rate": 0.0029076365685141463, + "loss": 5.801015377044678, + "step": 496 + }, + { + "epoch": 6.908296943231441, + "grad_norm": 0.008777810260653496, + "learning_rate": 0.00290717325393973, + "loss": 5.779858589172363, + "step": 497 + }, + { + "epoch": 6.922270742358078, + "grad_norm": 0.011423609219491482, + "learning_rate": 0.002906708817716469, + "loss": 5.777166843414307, + "step": 498 + }, + { + "epoch": 6.936244541484716, + "grad_norm": 0.008639856241643429, + "learning_rate": 0.00290624326021855, + "loss": 5.747189521789551, + "step": 499 + }, + { + "epoch": 6.950218340611354, + "grad_norm": 0.009628898464143276, + "learning_rate": 0.002905776581821066, + "loss": 5.7666916847229, + "step": 500 + }, + { + "epoch": 6.964192139737992, + "grad_norm": 0.011988377198576927, + "learning_rate": 0.0029053087829000097, + "loss": 5.777951240539551, + "step": 501 + }, + { + "epoch": 6.978165938864628, + "grad_norm": 0.01693224348127842, + "learning_rate": 0.00290483986383228, + "loss": 5.818393707275391, + "step": 502 + }, + { + "epoch": 6.992139737991266, + "grad_norm": 0.020264796912670135, + "learning_rate": 0.002904369824995674, + "loss": 5.747611045837402, + "step": 503 + }, + { + "epoch": 7.0, + "grad_norm": 0.018283747136592865, + "learning_rate": 0.0029038986667688964, + "loss": 5.789663314819336, + "step": 504 + }, + { + "epoch": 7.0, + "eval_loss": 5.825676918029785, + "eval_runtime": 57.6908, + "eval_samples_per_second": 42.329, + "eval_steps_per_second": 1.335, + "step": 504 + }, + { + "epoch": 7.013973799126638, + "grad_norm": 0.011447477154433727, + "learning_rate": 0.002903426389531549, + "loss": 5.724975109100342, + "step": 505 + }, + { + "epoch": 7.0279475982532755, + "grad_norm": 0.012539781630039215, + "learning_rate": 0.002902952993664138, + "loss": 5.747196674346924, + "step": 506 + }, + { + "epoch": 7.041921397379912, + "grad_norm": 0.012179837562143803, + "learning_rate": 0.002902478479548069, + "loss": 5.748995780944824, + "step": 507 + }, + { + "epoch": 7.05589519650655, + "grad_norm": 0.019896456971764565, + "learning_rate": 0.00290200284756565, + "loss": 5.751996994018555, + "step": 508 + }, + { + "epoch": 7.069868995633188, + "grad_norm": 0.022261742502450943, + "learning_rate": 0.002901526098100089, + "loss": 5.7516984939575195, + "step": 509 + }, + { + "epoch": 7.083842794759826, + "grad_norm": 0.02683536894619465, + "learning_rate": 0.0029010482315354946, + "loss": 5.716037750244141, + "step": 510 + }, + { + "epoch": 7.097816593886463, + "grad_norm": 0.018057072535157204, + "learning_rate": 0.0029005692482568747, + "loss": 5.778608322143555, + "step": 511 + }, + { + "epoch": 7.1117903930131, + "grad_norm": 0.011646891944110394, + "learning_rate": 0.002900089148650138, + "loss": 5.7308268547058105, + "step": 512 + }, + { + "epoch": 7.125764192139738, + "grad_norm": 0.02230495773255825, + "learning_rate": 0.002899607933102092, + "loss": 5.768552780151367, + "step": 513 + }, + { + "epoch": 7.139737991266376, + "grad_norm": 0.01212464738637209, + "learning_rate": 0.002899125602000443, + "loss": 5.733015060424805, + "step": 514 + }, + { + "epoch": 7.153711790393013, + "grad_norm": 0.011764907278120518, + "learning_rate": 0.0028986421557337973, + "loss": 5.748104095458984, + "step": 515 + }, + { + "epoch": 7.1676855895196505, + "grad_norm": 0.01593147963285446, + "learning_rate": 0.002898157594691658, + "loss": 5.701960563659668, + "step": 516 + }, + { + "epoch": 7.181659388646288, + "grad_norm": 0.016333773732185364, + "learning_rate": 0.0028976719192644277, + "loss": 5.716808795928955, + "step": 517 + }, + { + "epoch": 7.195633187772926, + "grad_norm": 0.017547868192195892, + "learning_rate": 0.002897185129843406, + "loss": 5.724377632141113, + "step": 518 + }, + { + "epoch": 7.209606986899563, + "grad_norm": 0.01662801019847393, + "learning_rate": 0.002896697226820791, + "loss": 5.737432956695557, + "step": 519 + }, + { + "epoch": 7.223580786026201, + "grad_norm": 0.015781691297888756, + "learning_rate": 0.002896208210589676, + "loss": 5.7376203536987305, + "step": 520 + }, + { + "epoch": 7.2375545851528384, + "grad_norm": 0.014363214373588562, + "learning_rate": 0.0028957180815440534, + "loss": 5.7342610359191895, + "step": 521 + }, + { + "epoch": 7.251528384279476, + "grad_norm": 0.015048054046928883, + "learning_rate": 0.0028952268400788115, + "loss": 5.715073585510254, + "step": 522 + }, + { + "epoch": 7.265502183406113, + "grad_norm": 0.013753226026892662, + "learning_rate": 0.002894734486589734, + "loss": 5.73264217376709, + "step": 523 + }, + { + "epoch": 7.279475982532751, + "grad_norm": 0.015562563203275204, + "learning_rate": 0.002894241021473502, + "loss": 5.7072038650512695, + "step": 524 + }, + { + "epoch": 7.293449781659389, + "grad_norm": 0.01520819216966629, + "learning_rate": 0.0028937464451276916, + "loss": 5.683300018310547, + "step": 525 + }, + { + "epoch": 7.307423580786026, + "grad_norm": 0.013024724088609219, + "learning_rate": 0.002893250757950773, + "loss": 5.726273536682129, + "step": 526 + }, + { + "epoch": 7.321397379912664, + "grad_norm": 0.012760856188833714, + "learning_rate": 0.0028927539603421128, + "loss": 5.7249064445495605, + "step": 527 + }, + { + "epoch": 7.335371179039301, + "grad_norm": 0.013542579486966133, + "learning_rate": 0.0028922560527019723, + "loss": 5.70849084854126, + "step": 528 + }, + { + "epoch": 7.349344978165939, + "grad_norm": 0.014801754616200924, + "learning_rate": 0.002891757035431506, + "loss": 5.725185394287109, + "step": 529 + }, + { + "epoch": 7.3633187772925766, + "grad_norm": 0.013000023551285267, + "learning_rate": 0.0028912569089327637, + "loss": 5.705799102783203, + "step": 530 + }, + { + "epoch": 7.377292576419214, + "grad_norm": 0.012919292785227299, + "learning_rate": 0.002890755673608688, + "loss": 5.674324989318848, + "step": 531 + }, + { + "epoch": 7.391266375545851, + "grad_norm": 0.015904201194643974, + "learning_rate": 0.002890253329863115, + "loss": 5.680788993835449, + "step": 532 + }, + { + "epoch": 7.405240174672489, + "grad_norm": 0.015467721968889236, + "learning_rate": 0.0028897498781007744, + "loss": 5.707531929016113, + "step": 533 + }, + { + "epoch": 7.419213973799127, + "grad_norm": 0.015157260000705719, + "learning_rate": 0.002889245318727288, + "loss": 5.689894676208496, + "step": 534 + }, + { + "epoch": 7.4331877729257645, + "grad_norm": 0.014204306527972221, + "learning_rate": 0.00288873965214917, + "loss": 5.713001251220703, + "step": 535 + }, + { + "epoch": 7.447161572052401, + "grad_norm": 0.022045845165848732, + "learning_rate": 0.0028882328787738273, + "loss": 5.657631874084473, + "step": 536 + }, + { + "epoch": 7.461135371179039, + "grad_norm": 0.013380175456404686, + "learning_rate": 0.0028877249990095573, + "loss": 5.702192306518555, + "step": 537 + }, + { + "epoch": 7.475109170305677, + "grad_norm": 0.016698408871889114, + "learning_rate": 0.0028872160132655503, + "loss": 5.695206642150879, + "step": 538 + }, + { + "epoch": 7.489082969432315, + "grad_norm": 0.020001472905278206, + "learning_rate": 0.0028867059219518866, + "loss": 5.731742858886719, + "step": 539 + }, + { + "epoch": 7.503056768558952, + "grad_norm": 0.027350103482604027, + "learning_rate": 0.0028861947254795375, + "loss": 5.712314605712891, + "step": 540 + }, + { + "epoch": 7.517030567685589, + "grad_norm": 0.021862946450710297, + "learning_rate": 0.002885682424260365, + "loss": 5.697043418884277, + "step": 541 + }, + { + "epoch": 7.531004366812227, + "grad_norm": 0.019216977059841156, + "learning_rate": 0.0028851690187071204, + "loss": 5.698813438415527, + "step": 542 + }, + { + "epoch": 7.544978165938865, + "grad_norm": 0.02131095714867115, + "learning_rate": 0.002884654509233446, + "loss": 5.6918864250183105, + "step": 543 + }, + { + "epoch": 7.558951965065502, + "grad_norm": 0.016583887860178947, + "learning_rate": 0.0028841388962538724, + "loss": 5.711456298828125, + "step": 544 + }, + { + "epoch": 7.5729257641921395, + "grad_norm": 0.016998684033751488, + "learning_rate": 0.0028836221801838194, + "loss": 5.675747871398926, + "step": 545 + }, + { + "epoch": 7.586899563318777, + "grad_norm": 0.01705755852162838, + "learning_rate": 0.0028831043614395966, + "loss": 5.697177410125732, + "step": 546 + }, + { + "epoch": 7.600873362445415, + "grad_norm": 0.014604352414608002, + "learning_rate": 0.0028825854404384, + "loss": 5.681950092315674, + "step": 547 + }, + { + "epoch": 7.614847161572053, + "grad_norm": 0.013561096973717213, + "learning_rate": 0.0028820654175983162, + "loss": 5.67357063293457, + "step": 548 + }, + { + "epoch": 7.62882096069869, + "grad_norm": 0.021149173378944397, + "learning_rate": 0.0028815442933383174, + "loss": 5.709477424621582, + "step": 549 + }, + { + "epoch": 7.642794759825327, + "grad_norm": 0.01431041955947876, + "learning_rate": 0.0028810220680782644, + "loss": 5.7012200355529785, + "step": 550 + }, + { + "epoch": 7.656768558951965, + "grad_norm": 0.01429709978401661, + "learning_rate": 0.0028804987422389045, + "loss": 5.689794540405273, + "step": 551 + }, + { + "epoch": 7.670742358078603, + "grad_norm": 0.032025713473558426, + "learning_rate": 0.0028799743162418712, + "loss": 5.6713762283325195, + "step": 552 + }, + { + "epoch": 7.68471615720524, + "grad_norm": 0.013168714940547943, + "learning_rate": 0.002879448790509686, + "loss": 5.685478210449219, + "step": 553 + }, + { + "epoch": 7.698689956331878, + "grad_norm": 0.014088666997849941, + "learning_rate": 0.0028789221654657556, + "loss": 5.645514965057373, + "step": 554 + }, + { + "epoch": 7.712663755458515, + "grad_norm": 0.013320053927600384, + "learning_rate": 0.002878394441534372, + "loss": 5.678678512573242, + "step": 555 + }, + { + "epoch": 7.726637554585153, + "grad_norm": 0.022180087864398956, + "learning_rate": 0.002877865619140712, + "loss": 5.660525321960449, + "step": 556 + }, + { + "epoch": 7.74061135371179, + "grad_norm": 0.01271889265626669, + "learning_rate": 0.002877335698710839, + "loss": 5.652164459228516, + "step": 557 + }, + { + "epoch": 7.754585152838428, + "grad_norm": 0.011617431417107582, + "learning_rate": 0.002876804680671701, + "loss": 5.6389946937561035, + "step": 558 + }, + { + "epoch": 7.7685589519650655, + "grad_norm": 0.012040726840496063, + "learning_rate": 0.0028762725654511282, + "loss": 5.628744602203369, + "step": 559 + }, + { + "epoch": 7.782532751091703, + "grad_norm": 0.013148069381713867, + "learning_rate": 0.0028757393534778377, + "loss": 5.69610595703125, + "step": 560 + }, + { + "epoch": 7.796506550218341, + "grad_norm": 0.01826276071369648, + "learning_rate": 0.0028752050451814276, + "loss": 5.634238243103027, + "step": 561 + }, + { + "epoch": 7.810480349344978, + "grad_norm": 0.020393460988998413, + "learning_rate": 0.0028746696409923805, + "loss": 5.656968593597412, + "step": 562 + }, + { + "epoch": 7.824454148471616, + "grad_norm": 0.017072148621082306, + "learning_rate": 0.0028741331413420627, + "loss": 5.668219089508057, + "step": 563 + }, + { + "epoch": 7.8384279475982535, + "grad_norm": 0.01694098301231861, + "learning_rate": 0.0028735955466627217, + "loss": 5.670073509216309, + "step": 564 + }, + { + "epoch": 7.85240174672489, + "grad_norm": 0.022760670632123947, + "learning_rate": 0.002873056857387488, + "loss": 5.649288177490234, + "step": 565 + }, + { + "epoch": 7.866375545851528, + "grad_norm": 0.03613442927598953, + "learning_rate": 0.0028725170739503734, + "loss": 5.723211288452148, + "step": 566 + }, + { + "epoch": 7.880349344978166, + "grad_norm": 0.04232329502701759, + "learning_rate": 0.002871976196786272, + "loss": 5.757920742034912, + "step": 567 + }, + { + "epoch": 7.894323144104804, + "grad_norm": 0.044210851192474365, + "learning_rate": 0.002871434226330959, + "loss": 5.702426910400391, + "step": 568 + }, + { + "epoch": 7.908296943231441, + "grad_norm": 0.044886402785778046, + "learning_rate": 0.0028708911630210905, + "loss": 5.742881774902344, + "step": 569 + }, + { + "epoch": 7.922270742358078, + "grad_norm": 0.022805195301771164, + "learning_rate": 0.002870347007294202, + "loss": 5.716353416442871, + "step": 570 + }, + { + "epoch": 7.936244541484716, + "grad_norm": 0.02219395525753498, + "learning_rate": 0.00286980175958871, + "loss": 5.68503475189209, + "step": 571 + }, + { + "epoch": 7.950218340611354, + "grad_norm": 0.019240057095885277, + "learning_rate": 0.002869255420343911, + "loss": 5.700656890869141, + "step": 572 + }, + { + "epoch": 7.964192139737992, + "grad_norm": 0.01957295648753643, + "learning_rate": 0.0028687079899999807, + "loss": 5.70595645904541, + "step": 573 + }, + { + "epoch": 7.978165938864628, + "grad_norm": 0.016326555982232094, + "learning_rate": 0.0028681594689979736, + "loss": 5.697229385375977, + "step": 574 + }, + { + "epoch": 7.992139737991266, + "grad_norm": 0.015815885737538338, + "learning_rate": 0.0028676098577798234, + "loss": 5.7020792961120605, + "step": 575 + }, + { + "epoch": 8.0, + "grad_norm": 0.015865175053477287, + "learning_rate": 0.002867059156788342, + "loss": 5.648272514343262, + "step": 576 + }, + { + "epoch": 8.0, + "eval_loss": 5.7417449951171875, + "eval_runtime": 51.2263, + "eval_samples_per_second": 47.671, + "eval_steps_per_second": 1.503, + "step": 576 + }, + { + "epoch": 8.013973799126637, + "grad_norm": 0.016063183546066284, + "learning_rate": 0.0028665073664672185, + "loss": 5.65411376953125, + "step": 577 + }, + { + "epoch": 8.027947598253276, + "grad_norm": 0.014323881827294827, + "learning_rate": 0.0028659544872610214, + "loss": 5.629649639129639, + "step": 578 + }, + { + "epoch": 8.041921397379912, + "grad_norm": 0.014005238190293312, + "learning_rate": 0.002865400519615195, + "loss": 5.655135631561279, + "step": 579 + }, + { + "epoch": 8.055895196506551, + "grad_norm": 0.01385372132062912, + "learning_rate": 0.0028648454639760615, + "loss": 5.632213115692139, + "step": 580 + }, + { + "epoch": 8.069868995633188, + "grad_norm": 0.014073756523430347, + "learning_rate": 0.0028642893207908187, + "loss": 5.626431465148926, + "step": 581 + }, + { + "epoch": 8.083842794759825, + "grad_norm": 0.019551943987607956, + "learning_rate": 0.0028637320905075406, + "loss": 5.595447063446045, + "step": 582 + }, + { + "epoch": 8.097816593886463, + "grad_norm": 0.017571093514561653, + "learning_rate": 0.002863173773575179, + "loss": 5.661581039428711, + "step": 583 + }, + { + "epoch": 8.1117903930131, + "grad_norm": 0.01691303588449955, + "learning_rate": 0.0028626143704435585, + "loss": 5.641129493713379, + "step": 584 + }, + { + "epoch": 8.125764192139737, + "grad_norm": 0.01584242656826973, + "learning_rate": 0.0028620538815633807, + "loss": 5.640626907348633, + "step": 585 + }, + { + "epoch": 8.139737991266376, + "grad_norm": 0.014592235907912254, + "learning_rate": 0.0028614923073862213, + "loss": 5.611204147338867, + "step": 586 + }, + { + "epoch": 8.153711790393013, + "grad_norm": 0.016636405140161514, + "learning_rate": 0.0028609296483645304, + "loss": 5.614716053009033, + "step": 587 + }, + { + "epoch": 8.167685589519651, + "grad_norm": 0.01370833721011877, + "learning_rate": 0.002860365904951632, + "loss": 5.633011817932129, + "step": 588 + }, + { + "epoch": 8.181659388646288, + "grad_norm": 0.013218311592936516, + "learning_rate": 0.002859801077601725, + "loss": 5.588958740234375, + "step": 589 + }, + { + "epoch": 8.195633187772925, + "grad_norm": 0.016971616074442863, + "learning_rate": 0.0028592351667698786, + "loss": 5.635064125061035, + "step": 590 + }, + { + "epoch": 8.209606986899564, + "grad_norm": 0.015524905174970627, + "learning_rate": 0.0028586681729120393, + "loss": 5.611454963684082, + "step": 591 + }, + { + "epoch": 8.2235807860262, + "grad_norm": 0.014769157394766808, + "learning_rate": 0.0028581000964850215, + "loss": 5.62175178527832, + "step": 592 + }, + { + "epoch": 8.237554585152838, + "grad_norm": 0.014313343912363052, + "learning_rate": 0.002857530937946516, + "loss": 5.603628158569336, + "step": 593 + }, + { + "epoch": 8.251528384279476, + "grad_norm": 0.016748322173953056, + "learning_rate": 0.002856960697755083, + "loss": 5.617765426635742, + "step": 594 + }, + { + "epoch": 8.265502183406113, + "grad_norm": 0.013503300957381725, + "learning_rate": 0.0028563893763701538, + "loss": 5.636135578155518, + "step": 595 + }, + { + "epoch": 8.279475982532752, + "grad_norm": 0.012275655753910542, + "learning_rate": 0.002855816974252033, + "loss": 5.552533149719238, + "step": 596 + }, + { + "epoch": 8.293449781659389, + "grad_norm": 0.01288773212581873, + "learning_rate": 0.002855243491861894, + "loss": 5.620527267456055, + "step": 597 + }, + { + "epoch": 8.307423580786025, + "grad_norm": 0.013687008060514927, + "learning_rate": 0.002854668929661781, + "loss": 5.618595123291016, + "step": 598 + }, + { + "epoch": 8.321397379912664, + "grad_norm": 0.015678439289331436, + "learning_rate": 0.0028540932881146092, + "loss": 5.604031562805176, + "step": 599 + }, + { + "epoch": 8.335371179039301, + "grad_norm": 0.013567988760769367, + "learning_rate": 0.002853516567684162, + "loss": 5.549149513244629, + "step": 600 + }, + { + "epoch": 8.34934497816594, + "grad_norm": 0.0157297495752573, + "learning_rate": 0.002852938768835092, + "loss": 5.618505477905273, + "step": 601 + }, + { + "epoch": 8.363318777292577, + "grad_norm": 0.011867019347846508, + "learning_rate": 0.0028523598920329233, + "loss": 5.579049110412598, + "step": 602 + }, + { + "epoch": 8.377292576419213, + "grad_norm": 0.01398563478142023, + "learning_rate": 0.0028517799377440443, + "loss": 5.597678184509277, + "step": 603 + }, + { + "epoch": 8.391266375545852, + "grad_norm": 0.015525273978710175, + "learning_rate": 0.0028511989064357148, + "loss": 5.601663589477539, + "step": 604 + }, + { + "epoch": 8.405240174672489, + "grad_norm": 0.01588466763496399, + "learning_rate": 0.002850616798576061, + "loss": 5.5650434494018555, + "step": 605 + }, + { + "epoch": 8.419213973799126, + "grad_norm": 0.014276969246566296, + "learning_rate": 0.0028500336146340772, + "loss": 5.607125282287598, + "step": 606 + }, + { + "epoch": 8.433187772925764, + "grad_norm": 0.013253689743578434, + "learning_rate": 0.002849449355079623, + "loss": 5.6299824714660645, + "step": 607 + }, + { + "epoch": 8.447161572052401, + "grad_norm": 0.01176661066710949, + "learning_rate": 0.0028488640203834263, + "loss": 5.560839653015137, + "step": 608 + }, + { + "epoch": 8.46113537117904, + "grad_norm": 0.011712565086781979, + "learning_rate": 0.0028482776110170813, + "loss": 5.618312835693359, + "step": 609 + }, + { + "epoch": 8.475109170305677, + "grad_norm": 0.013515647500753403, + "learning_rate": 0.002847690127453047, + "loss": 5.566764831542969, + "step": 610 + }, + { + "epoch": 8.489082969432314, + "grad_norm": 0.020177524536848068, + "learning_rate": 0.0028471015701646485, + "loss": 5.554571628570557, + "step": 611 + }, + { + "epoch": 8.503056768558952, + "grad_norm": 0.029197178781032562, + "learning_rate": 0.0028465119396260757, + "loss": 5.6097259521484375, + "step": 612 + }, + { + "epoch": 8.51703056768559, + "grad_norm": 0.027947140857577324, + "learning_rate": 0.0028459212363123824, + "loss": 5.65598201751709, + "step": 613 + }, + { + "epoch": 8.531004366812226, + "grad_norm": 0.021089809015393257, + "learning_rate": 0.0028453294606994887, + "loss": 5.67195987701416, + "step": 614 + }, + { + "epoch": 8.544978165938865, + "grad_norm": 0.019936271011829376, + "learning_rate": 0.0028447366132641773, + "loss": 5.609482765197754, + "step": 615 + }, + { + "epoch": 8.558951965065502, + "grad_norm": 0.0176444910466671, + "learning_rate": 0.002844142694484094, + "loss": 5.60432767868042, + "step": 616 + }, + { + "epoch": 8.57292576419214, + "grad_norm": 0.023154428228735924, + "learning_rate": 0.0028435477048377495, + "loss": 5.616270065307617, + "step": 617 + }, + { + "epoch": 8.586899563318777, + "grad_norm": 0.018391480669379234, + "learning_rate": 0.002842951644804515, + "loss": 5.600588321685791, + "step": 618 + }, + { + "epoch": 8.600873362445414, + "grad_norm": 0.013716557994484901, + "learning_rate": 0.002842354514864627, + "loss": 5.595095634460449, + "step": 619 + }, + { + "epoch": 8.614847161572053, + "grad_norm": 0.014372860081493855, + "learning_rate": 0.002841756315499181, + "loss": 5.584517478942871, + "step": 620 + }, + { + "epoch": 8.62882096069869, + "grad_norm": 0.014930845238268375, + "learning_rate": 0.002841157047190135, + "loss": 5.587005138397217, + "step": 621 + }, + { + "epoch": 8.642794759825328, + "grad_norm": 0.012631502002477646, + "learning_rate": 0.0028405567104203095, + "loss": 5.594094276428223, + "step": 622 + }, + { + "epoch": 8.656768558951965, + "grad_norm": 0.014677002094686031, + "learning_rate": 0.0028399553056733854, + "loss": 5.569582939147949, + "step": 623 + }, + { + "epoch": 8.670742358078602, + "grad_norm": 0.015558961778879166, + "learning_rate": 0.0028393528334339025, + "loss": 5.596535682678223, + "step": 624 + }, + { + "epoch": 8.68471615720524, + "grad_norm": 0.014322095550596714, + "learning_rate": 0.0028387492941872624, + "loss": 5.570324420928955, + "step": 625 + }, + { + "epoch": 8.698689956331878, + "grad_norm": 0.012913414277136326, + "learning_rate": 0.002838144688419726, + "loss": 5.573002338409424, + "step": 626 + }, + { + "epoch": 8.712663755458514, + "grad_norm": 0.01251310296356678, + "learning_rate": 0.0028375390166184137, + "loss": 5.550327777862549, + "step": 627 + }, + { + "epoch": 8.726637554585153, + "grad_norm": 0.013847017660737038, + "learning_rate": 0.0028369322792713026, + "loss": 5.54567813873291, + "step": 628 + }, + { + "epoch": 8.74061135371179, + "grad_norm": 0.013789358548820019, + "learning_rate": 0.0028363244768672313, + "loss": 5.571495056152344, + "step": 629 + }, + { + "epoch": 8.754585152838429, + "grad_norm": 0.01236631814390421, + "learning_rate": 0.0028357156098958954, + "loss": 5.569730758666992, + "step": 630 + }, + { + "epoch": 8.768558951965066, + "grad_norm": 0.013210169039666653, + "learning_rate": 0.002835105678847847, + "loss": 5.53984260559082, + "step": 631 + }, + { + "epoch": 8.782532751091702, + "grad_norm": 0.01604137383401394, + "learning_rate": 0.0028344946842144975, + "loss": 5.530598163604736, + "step": 632 + }, + { + "epoch": 8.796506550218341, + "grad_norm": 0.01708815060555935, + "learning_rate": 0.002833882626488114, + "loss": 5.529256820678711, + "step": 633 + }, + { + "epoch": 8.810480349344978, + "grad_norm": 0.017222166061401367, + "learning_rate": 0.0028332695061618197, + "loss": 5.564949035644531, + "step": 634 + }, + { + "epoch": 8.824454148471617, + "grad_norm": 0.01803218014538288, + "learning_rate": 0.002832655323729595, + "loss": 5.560049057006836, + "step": 635 + }, + { + "epoch": 8.838427947598253, + "grad_norm": 0.01658225618302822, + "learning_rate": 0.002832040079686276, + "loss": 5.553183555603027, + "step": 636 + }, + { + "epoch": 8.85240174672489, + "grad_norm": 0.016834143549203873, + "learning_rate": 0.002831423774527554, + "loss": 5.573678970336914, + "step": 637 + }, + { + "epoch": 8.866375545851529, + "grad_norm": 0.01856408454477787, + "learning_rate": 0.002830806408749974, + "loss": 5.539497375488281, + "step": 638 + }, + { + "epoch": 8.880349344978166, + "grad_norm": 0.026909375563263893, + "learning_rate": 0.002830187982850936, + "loss": 5.551433563232422, + "step": 639 + }, + { + "epoch": 8.894323144104803, + "grad_norm": 0.016046348959207535, + "learning_rate": 0.002829568497328696, + "loss": 5.566728591918945, + "step": 640 + }, + { + "epoch": 8.908296943231441, + "grad_norm": 0.018582617864012718, + "learning_rate": 0.0028289479526823614, + "loss": 5.543226718902588, + "step": 641 + }, + { + "epoch": 8.922270742358078, + "grad_norm": 0.0211537666618824, + "learning_rate": 0.0028283263494118943, + "loss": 5.560711860656738, + "step": 642 + }, + { + "epoch": 8.936244541484717, + "grad_norm": 0.013894101604819298, + "learning_rate": 0.0028277036880181095, + "loss": 5.528890132904053, + "step": 643 + }, + { + "epoch": 8.950218340611354, + "grad_norm": 0.017581885680556297, + "learning_rate": 0.0028270799690026725, + "loss": 5.57473087310791, + "step": 644 + }, + { + "epoch": 8.96419213973799, + "grad_norm": 0.01622696965932846, + "learning_rate": 0.002826455192868104, + "loss": 5.578961372375488, + "step": 645 + }, + { + "epoch": 8.97816593886463, + "grad_norm": 0.01966337114572525, + "learning_rate": 0.0028258293601177743, + "loss": 5.513942718505859, + "step": 646 + }, + { + "epoch": 8.992139737991266, + "grad_norm": 0.021581366658210754, + "learning_rate": 0.002825202471255906, + "loss": 5.573669910430908, + "step": 647 + }, + { + "epoch": 9.0, + "grad_norm": 0.020468758419156075, + "learning_rate": 0.002824574526787571, + "loss": 5.516103744506836, + "step": 648 + }, + { + "epoch": 9.0, + "eval_loss": 5.626922130584717, + "eval_runtime": 54.2404, + "eval_samples_per_second": 45.022, + "eval_steps_per_second": 1.42, + "step": 648 + }, + { + "epoch": 9.013973799126637, + "grad_norm": 0.014597049914300442, + "learning_rate": 0.0028239455272186936, + "loss": 5.506453514099121, + "step": 649 + }, + { + "epoch": 9.027947598253276, + "grad_norm": 0.020277054980397224, + "learning_rate": 0.002823315473056048, + "loss": 5.535350799560547, + "step": 650 + }, + { + "epoch": 9.041921397379912, + "grad_norm": 0.019103925675153732, + "learning_rate": 0.0028226843648072562, + "loss": 5.522078037261963, + "step": 651 + }, + { + "epoch": 9.055895196506551, + "grad_norm": 0.07158710062503815, + "learning_rate": 0.0028220522029807913, + "loss": 5.531142234802246, + "step": 652 + }, + { + "epoch": 9.069868995633188, + "grad_norm": 0.01958007924258709, + "learning_rate": 0.0028214189880859746, + "loss": 5.484055042266846, + "step": 653 + }, + { + "epoch": 9.083842794759825, + "grad_norm": 0.0185184795409441, + "learning_rate": 0.002820784720632976, + "loss": 5.511264801025391, + "step": 654 + }, + { + "epoch": 9.097816593886463, + "grad_norm": 0.01976991444826126, + "learning_rate": 0.0028201494011328133, + "loss": 5.51821231842041, + "step": 655 + }, + { + "epoch": 9.1117903930131, + "grad_norm": 0.02188604325056076, + "learning_rate": 0.002819513030097352, + "loss": 5.483095169067383, + "step": 656 + }, + { + "epoch": 9.125764192139737, + "grad_norm": 0.019405223429203033, + "learning_rate": 0.002818875608039305, + "loss": 5.539701461791992, + "step": 657 + }, + { + "epoch": 9.139737991266376, + "grad_norm": 0.016907615587115288, + "learning_rate": 0.0028182371354722306, + "loss": 5.518251419067383, + "step": 658 + }, + { + "epoch": 9.153711790393013, + "grad_norm": 0.02160538360476494, + "learning_rate": 0.0028175976129105364, + "loss": 5.509786605834961, + "step": 659 + }, + { + "epoch": 9.167685589519651, + "grad_norm": 0.023125721141695976, + "learning_rate": 0.002816957040869473, + "loss": 5.535561561584473, + "step": 660 + }, + { + "epoch": 9.181659388646288, + "grad_norm": 0.022233380004763603, + "learning_rate": 0.0028163154198651372, + "loss": 5.553831100463867, + "step": 661 + }, + { + "epoch": 9.195633187772925, + "grad_norm": 0.017227552831172943, + "learning_rate": 0.002815672750414473, + "loss": 5.547080993652344, + "step": 662 + }, + { + "epoch": 9.209606986899564, + "grad_norm": 0.017122473567724228, + "learning_rate": 0.0028150290330352665, + "loss": 5.493778228759766, + "step": 663 + }, + { + "epoch": 9.2235807860262, + "grad_norm": 0.020641913637518883, + "learning_rate": 0.0028143842682461498, + "loss": 5.536645889282227, + "step": 664 + }, + { + "epoch": 9.237554585152838, + "grad_norm": 0.017383966594934464, + "learning_rate": 0.0028137384565665967, + "loss": 5.506233215332031, + "step": 665 + }, + { + "epoch": 9.251528384279476, + "grad_norm": 0.017969606444239616, + "learning_rate": 0.002813091598516928, + "loss": 5.524741172790527, + "step": 666 + }, + { + "epoch": 9.265502183406113, + "grad_norm": 0.01936613954603672, + "learning_rate": 0.0028124436946183047, + "loss": 5.460989952087402, + "step": 667 + }, + { + "epoch": 9.279475982532752, + "grad_norm": 0.017492495477199554, + "learning_rate": 0.0028117947453927313, + "loss": 5.487619876861572, + "step": 668 + }, + { + "epoch": 9.293449781659389, + "grad_norm": 0.018786396831274033, + "learning_rate": 0.0028111447513630545, + "loss": 5.516829013824463, + "step": 669 + }, + { + "epoch": 9.307423580786025, + "grad_norm": 0.020149117335677147, + "learning_rate": 0.0028104937130529624, + "loss": 5.460541725158691, + "step": 670 + }, + { + "epoch": 9.321397379912664, + "grad_norm": 0.024668103083968163, + "learning_rate": 0.0028098416309869857, + "loss": 5.497687339782715, + "step": 671 + }, + { + "epoch": 9.335371179039301, + "grad_norm": 0.015888608992099762, + "learning_rate": 0.002809188505690494, + "loss": 5.494459629058838, + "step": 672 + }, + { + "epoch": 9.34934497816594, + "grad_norm": 0.019327623769640923, + "learning_rate": 0.0028085343376896997, + "loss": 5.486351490020752, + "step": 673 + }, + { + "epoch": 9.363318777292577, + "grad_norm": 0.026855193078517914, + "learning_rate": 0.0028078791275116532, + "loss": 5.509825706481934, + "step": 674 + }, + { + "epoch": 9.377292576419213, + "grad_norm": 0.026340370997786522, + "learning_rate": 0.002807222875684246, + "loss": 5.527161598205566, + "step": 675 + }, + { + "epoch": 9.391266375545852, + "grad_norm": 0.02200835384428501, + "learning_rate": 0.002806565582736208, + "loss": 5.507146835327148, + "step": 676 + }, + { + "epoch": 9.405240174672489, + "grad_norm": 0.020420534536242485, + "learning_rate": 0.0028059072491971085, + "loss": 5.5162506103515625, + "step": 677 + }, + { + "epoch": 9.419213973799126, + "grad_norm": 0.02493203990161419, + "learning_rate": 0.0028052478755973545, + "loss": 5.5252556800842285, + "step": 678 + }, + { + "epoch": 9.433187772925764, + "grad_norm": 0.021293973550200462, + "learning_rate": 0.0028045874624681925, + "loss": 5.523589134216309, + "step": 679 + }, + { + "epoch": 9.447161572052401, + "grad_norm": 0.016593236476182938, + "learning_rate": 0.0028039260103417045, + "loss": 5.500586032867432, + "step": 680 + }, + { + "epoch": 9.46113537117904, + "grad_norm": 0.018857723101973534, + "learning_rate": 0.002803263519750811, + "loss": 5.52933406829834, + "step": 681 + }, + { + "epoch": 9.475109170305677, + "grad_norm": 0.015930436551570892, + "learning_rate": 0.0028025999912292677, + "loss": 5.48813533782959, + "step": 682 + }, + { + "epoch": 9.489082969432314, + "grad_norm": 0.019427936524152756, + "learning_rate": 0.002801935425311669, + "loss": 5.46573543548584, + "step": 683 + }, + { + "epoch": 9.503056768558952, + "grad_norm": 0.018123053014278412, + "learning_rate": 0.0028012698225334427, + "loss": 5.499445915222168, + "step": 684 + }, + { + "epoch": 9.51703056768559, + "grad_norm": 0.02035410702228546, + "learning_rate": 0.002800603183430853, + "loss": 5.513540267944336, + "step": 685 + }, + { + "epoch": 9.531004366812226, + "grad_norm": 0.025838350877165794, + "learning_rate": 0.0027999355085409996, + "loss": 5.565637588500977, + "step": 686 + }, + { + "epoch": 9.544978165938865, + "grad_norm": 0.029621612280607224, + "learning_rate": 0.002799266798401816, + "loss": 5.5619215965271, + "step": 687 + }, + { + "epoch": 9.558951965065502, + "grad_norm": 0.028948379680514336, + "learning_rate": 0.0027985970535520685, + "loss": 5.551331520080566, + "step": 688 + }, + { + "epoch": 9.57292576419214, + "grad_norm": 0.022135930135846138, + "learning_rate": 0.0027979262745313604, + "loss": 5.527722358703613, + "step": 689 + }, + { + "epoch": 9.586899563318777, + "grad_norm": 0.028884069994091988, + "learning_rate": 0.0027972544618801256, + "loss": 5.554599761962891, + "step": 690 + }, + { + "epoch": 9.600873362445414, + "grad_norm": 0.02499174140393734, + "learning_rate": 0.00279658161613963, + "loss": 5.543351173400879, + "step": 691 + }, + { + "epoch": 9.614847161572053, + "grad_norm": 0.021131042391061783, + "learning_rate": 0.0027959077378519755, + "loss": 5.496039390563965, + "step": 692 + }, + { + "epoch": 9.62882096069869, + "grad_norm": 0.02957489714026451, + "learning_rate": 0.002795232827560092, + "loss": 5.552300453186035, + "step": 693 + }, + { + "epoch": 9.642794759825328, + "grad_norm": 0.022832119837403297, + "learning_rate": 0.0027945568858077433, + "loss": 5.488651275634766, + "step": 694 + }, + { + "epoch": 9.656768558951965, + "grad_norm": 0.02074701525270939, + "learning_rate": 0.002793879913139523, + "loss": 5.491244792938232, + "step": 695 + }, + { + "epoch": 9.670742358078602, + "grad_norm": 0.019919803366065025, + "learning_rate": 0.002793201910100856, + "loss": 5.483855247497559, + "step": 696 + }, + { + "epoch": 9.68471615720524, + "grad_norm": 0.01858169585466385, + "learning_rate": 0.002792522877237998, + "loss": 5.50787353515625, + "step": 697 + }, + { + "epoch": 9.698689956331878, + "grad_norm": 0.02085857465863228, + "learning_rate": 0.002791842815098032, + "loss": 5.512872695922852, + "step": 698 + }, + { + "epoch": 9.712663755458514, + "grad_norm": 0.01407450158149004, + "learning_rate": 0.002791161724228872, + "loss": 5.472590446472168, + "step": 699 + }, + { + "epoch": 9.726637554585153, + "grad_norm": 0.01672944612801075, + "learning_rate": 0.002790479605179262, + "loss": 5.469183921813965, + "step": 700 + }, + { + "epoch": 9.74061135371179, + "grad_norm": 0.016565700992941856, + "learning_rate": 0.0027897964584987713, + "loss": 5.493677139282227, + "step": 701 + }, + { + "epoch": 9.754585152838429, + "grad_norm": 0.017017360776662827, + "learning_rate": 0.0027891122847377996, + "loss": 5.464931488037109, + "step": 702 + }, + { + "epoch": 9.768558951965066, + "grad_norm": 0.019842486828565598, + "learning_rate": 0.002788427084447574, + "loss": 5.461293697357178, + "step": 703 + }, + { + "epoch": 9.782532751091702, + "grad_norm": 0.016210904344916344, + "learning_rate": 0.0027877408581801467, + "loss": 5.454716682434082, + "step": 704 + }, + { + "epoch": 9.796506550218341, + "grad_norm": 0.02064039371907711, + "learning_rate": 0.002787053606488398, + "loss": 5.466576099395752, + "step": 705 + }, + { + "epoch": 9.810480349344978, + "grad_norm": 0.018198711797595024, + "learning_rate": 0.0027863653299260356, + "loss": 5.447747230529785, + "step": 706 + }, + { + "epoch": 9.824454148471617, + "grad_norm": 0.015937626361846924, + "learning_rate": 0.00278567602904759, + "loss": 5.5055036544799805, + "step": 707 + }, + { + "epoch": 9.838427947598253, + "grad_norm": 0.01687259040772915, + "learning_rate": 0.002784985704408419, + "loss": 5.422250747680664, + "step": 708 + }, + { + "epoch": 9.85240174672489, + "grad_norm": 0.02622046135365963, + "learning_rate": 0.002784294356564705, + "loss": 5.466894626617432, + "step": 709 + }, + { + "epoch": 9.866375545851529, + "grad_norm": 0.018672186881303787, + "learning_rate": 0.0027836019860734535, + "loss": 5.48300838470459, + "step": 710 + }, + { + "epoch": 9.880349344978166, + "grad_norm": 0.02166241593658924, + "learning_rate": 0.0027829085934924958, + "loss": 5.477658271789551, + "step": 711 + }, + { + "epoch": 9.894323144104803, + "grad_norm": 0.018426023423671722, + "learning_rate": 0.0027822141793804856, + "loss": 5.487186431884766, + "step": 712 + }, + { + "epoch": 9.908296943231441, + "grad_norm": 0.022229891270399094, + "learning_rate": 0.0027815187442968995, + "loss": 5.488579273223877, + "step": 713 + }, + { + "epoch": 9.922270742358078, + "grad_norm": 0.026930972933769226, + "learning_rate": 0.002780822288802037, + "loss": 5.492000102996826, + "step": 714 + }, + { + "epoch": 9.936244541484717, + "grad_norm": 0.02733265422284603, + "learning_rate": 0.00278012481345702, + "loss": 5.4663801193237305, + "step": 715 + }, + { + "epoch": 9.950218340611354, + "grad_norm": 0.02993253618478775, + "learning_rate": 0.0027794263188237922, + "loss": 5.543368816375732, + "step": 716 + }, + { + "epoch": 9.96419213973799, + "grad_norm": 0.03152153640985489, + "learning_rate": 0.002778726805465118, + "loss": 5.565644264221191, + "step": 717 + }, + { + "epoch": 9.97816593886463, + "grad_norm": 0.031395044177770615, + "learning_rate": 0.002778026273944581, + "loss": 5.553359508514404, + "step": 718 + }, + { + "epoch": 9.992139737991266, + "grad_norm": 0.026603225618600845, + "learning_rate": 0.002777324724826589, + "loss": 5.552347183227539, + "step": 719 + }, + { + "epoch": 10.0, + "grad_norm": 0.024152139201760292, + "learning_rate": 0.002776622158676366, + "loss": 5.493257522583008, + "step": 720 + } + ], + "logging_steps": 1, + "max_steps": 3600, + "num_input_tokens_seen": 0, + "num_train_epochs": 50, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": true, + "should_log": false, + "should_save": false, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 3.069613361332224e+17, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +} diff --git a/runs/baseline/checkpoint-720/training_args.bin b/runs/baseline/checkpoint-720/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..621fd4a3c3234a9460e6640b7932c67b4dbaf6f7 --- /dev/null +++ b/runs/baseline/checkpoint-720/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f82546ea14de1f78b27ceb41ab9a039dadb9a2a47d76f51181545f79d58b9d97 +size 5265